00:00:00.000 Started by upstream project "autotest-spdk-master-vs-dpdk-v22.11" build number 2440 00:00:00.000 originally caused by: 00:00:00.001 Started by upstream project "nightly-trigger" build number 3705 00:00:00.001 originally caused by: 00:00:00.001 Started by timer 00:00:00.067 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:00.068 The recommended git tool is: git 00:00:00.068 using credential 00000000-0000-0000-0000-000000000002 00:00:00.070 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.103 Fetching changes from the remote Git repository 00:00:00.108 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.156 Using shallow fetch with depth 1 00:00:00.156 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.156 > git --version # timeout=10 00:00:00.206 > git --version # 'git version 2.39.2' 00:00:00.206 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.250 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.250 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:04.266 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:04.280 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:04.291 Checking out Revision db4637e8b949f278f369ec13f70585206ccd9507 (FETCH_HEAD) 00:00:04.291 > git config core.sparsecheckout # timeout=10 00:00:04.302 > git read-tree -mu HEAD # timeout=10 00:00:04.317 > git checkout -f db4637e8b949f278f369ec13f70585206ccd9507 # timeout=5 00:00:04.338 Commit message: "jenkins/jjb-config: Add missing SPDK_TEST_NVME_INTERRUPT flag" 00:00:04.339 > git rev-list --no-walk db4637e8b949f278f369ec13f70585206ccd9507 # timeout=10 00:00:04.426 [Pipeline] Start of Pipeline 00:00:04.440 [Pipeline] library 00:00:04.441 Loading library shm_lib@master 00:00:04.441 Library shm_lib@master is cached. Copying from home. 00:00:04.455 [Pipeline] node 00:00:04.477 Running on VM-host-WFP7 in /var/jenkins/workspace/raid-vg-autotest 00:00:04.479 [Pipeline] { 00:00:04.488 [Pipeline] catchError 00:00:04.490 [Pipeline] { 00:00:04.499 [Pipeline] wrap 00:00:04.506 [Pipeline] { 00:00:04.515 [Pipeline] stage 00:00:04.517 [Pipeline] { (Prologue) 00:00:04.537 [Pipeline] echo 00:00:04.539 Node: VM-host-WFP7 00:00:04.547 [Pipeline] cleanWs 00:00:04.559 [WS-CLEANUP] Deleting project workspace... 00:00:04.559 [WS-CLEANUP] Deferred wipeout is used... 00:00:04.567 [WS-CLEANUP] done 00:00:04.792 [Pipeline] setCustomBuildProperty 00:00:04.898 [Pipeline] httpRequest 00:00:05.307 [Pipeline] echo 00:00:05.309 Sorcerer 10.211.164.101 is alive 00:00:05.321 [Pipeline] retry 00:00:05.323 [Pipeline] { 00:00:05.337 [Pipeline] httpRequest 00:00:05.342 HttpMethod: GET 00:00:05.342 URL: http://10.211.164.101/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:05.343 Sending request to url: http://10.211.164.101/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:05.346 Response Code: HTTP/1.1 200 OK 00:00:05.346 Success: Status code 200 is in the accepted range: 200,404 00:00:05.347 Saving response body to /var/jenkins/workspace/raid-vg-autotest/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:05.674 [Pipeline] } 00:00:05.687 [Pipeline] // retry 00:00:05.694 [Pipeline] sh 00:00:05.979 + tar --no-same-owner -xf jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:05.994 [Pipeline] httpRequest 00:00:06.751 [Pipeline] echo 00:00:06.753 Sorcerer 10.211.164.101 is alive 00:00:06.761 [Pipeline] retry 00:00:06.763 [Pipeline] { 00:00:06.774 [Pipeline] httpRequest 00:00:06.779 HttpMethod: GET 00:00:06.779 URL: http://10.211.164.101/packages/spdk_a2f5e1c2d535934bced849d8b079523bc74c98f1.tar.gz 00:00:06.779 Sending request to url: http://10.211.164.101/packages/spdk_a2f5e1c2d535934bced849d8b079523bc74c98f1.tar.gz 00:00:06.783 Response Code: HTTP/1.1 200 OK 00:00:06.784 Success: Status code 200 is in the accepted range: 200,404 00:00:06.784 Saving response body to /var/jenkins/workspace/raid-vg-autotest/spdk_a2f5e1c2d535934bced849d8b079523bc74c98f1.tar.gz 00:00:28.293 [Pipeline] } 00:00:28.306 [Pipeline] // retry 00:00:28.312 [Pipeline] sh 00:00:28.594 + tar --no-same-owner -xf spdk_a2f5e1c2d535934bced849d8b079523bc74c98f1.tar.gz 00:00:31.152 [Pipeline] sh 00:00:31.438 + git -C spdk log --oneline -n5 00:00:31.438 a2f5e1c2d blob: don't free bs when spdk_bs_destroy/spdk_bs_unload fails 00:00:31.438 0f59982b6 blob: don't use bs_load_ctx_fail in bs_write_used_* functions 00:00:31.438 0354bb8e8 nvme/rdma: Force qp disconnect on pg remove 00:00:31.438 0ea9ac02f accel/mlx5: Create pool of UMRs 00:00:31.438 60adca7e1 lib/mlx5: API to configure UMR 00:00:31.462 [Pipeline] withCredentials 00:00:31.474 > git --version # timeout=10 00:00:31.487 > git --version # 'git version 2.39.2' 00:00:31.506 Masking supported pattern matches of $GIT_PASSWORD or $GIT_ASKPASS 00:00:31.509 [Pipeline] { 00:00:31.518 [Pipeline] retry 00:00:31.520 [Pipeline] { 00:00:31.535 [Pipeline] sh 00:00:31.820 + git ls-remote http://dpdk.org/git/dpdk-stable v22.11.4 00:00:32.776 [Pipeline] } 00:00:32.796 [Pipeline] // retry 00:00:32.802 [Pipeline] } 00:00:32.818 [Pipeline] // withCredentials 00:00:32.829 [Pipeline] httpRequest 00:00:33.254 [Pipeline] echo 00:00:33.256 Sorcerer 10.211.164.101 is alive 00:00:33.266 [Pipeline] retry 00:00:33.268 [Pipeline] { 00:00:33.282 [Pipeline] httpRequest 00:00:33.287 HttpMethod: GET 00:00:33.288 URL: http://10.211.164.101/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:00:33.288 Sending request to url: http://10.211.164.101/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:00:33.303 Response Code: HTTP/1.1 200 OK 00:00:33.304 Success: Status code 200 is in the accepted range: 200,404 00:00:33.304 Saving response body to /var/jenkins/workspace/raid-vg-autotest/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:01:10.219 [Pipeline] } 00:01:10.252 [Pipeline] // retry 00:01:10.258 [Pipeline] sh 00:01:10.539 + tar --no-same-owner -xf dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:01:11.933 [Pipeline] sh 00:01:12.214 + git -C dpdk log --oneline -n5 00:01:12.214 caf0f5d395 version: 22.11.4 00:01:12.214 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:01:12.214 dc9c799c7d vhost: fix missing spinlock unlock 00:01:12.214 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:01:12.214 6ef77f2a5e net/gve: fix RX buffer size alignment 00:01:12.231 [Pipeline] writeFile 00:01:12.246 [Pipeline] sh 00:01:12.530 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:01:12.543 [Pipeline] sh 00:01:12.833 + cat autorun-spdk.conf 00:01:12.833 SPDK_RUN_FUNCTIONAL_TEST=1 00:01:12.833 SPDK_RUN_ASAN=1 00:01:12.833 SPDK_RUN_UBSAN=1 00:01:12.833 SPDK_TEST_RAID=1 00:01:12.833 SPDK_TEST_NATIVE_DPDK=v22.11.4 00:01:12.833 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:12.833 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:12.839 RUN_NIGHTLY=1 00:01:12.841 [Pipeline] } 00:01:12.851 [Pipeline] // stage 00:01:12.862 [Pipeline] stage 00:01:12.864 [Pipeline] { (Run VM) 00:01:12.875 [Pipeline] sh 00:01:13.160 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:01:13.160 + echo 'Start stage prepare_nvme.sh' 00:01:13.160 Start stage prepare_nvme.sh 00:01:13.160 + [[ -n 1 ]] 00:01:13.160 + disk_prefix=ex1 00:01:13.160 + [[ -n /var/jenkins/workspace/raid-vg-autotest ]] 00:01:13.160 + [[ -e /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf ]] 00:01:13.160 + source /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf 00:01:13.160 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:13.160 ++ SPDK_RUN_ASAN=1 00:01:13.160 ++ SPDK_RUN_UBSAN=1 00:01:13.160 ++ SPDK_TEST_RAID=1 00:01:13.160 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:01:13.160 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:13.160 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:13.160 ++ RUN_NIGHTLY=1 00:01:13.160 + cd /var/jenkins/workspace/raid-vg-autotest 00:01:13.160 + nvme_files=() 00:01:13.160 + declare -A nvme_files 00:01:13.160 + backend_dir=/var/lib/libvirt/images/backends 00:01:13.160 + nvme_files['nvme.img']=5G 00:01:13.160 + nvme_files['nvme-cmb.img']=5G 00:01:13.160 + nvme_files['nvme-multi0.img']=4G 00:01:13.160 + nvme_files['nvme-multi1.img']=4G 00:01:13.160 + nvme_files['nvme-multi2.img']=4G 00:01:13.160 + nvme_files['nvme-openstack.img']=8G 00:01:13.160 + nvme_files['nvme-zns.img']=5G 00:01:13.160 + (( SPDK_TEST_NVME_PMR == 1 )) 00:01:13.160 + (( SPDK_TEST_FTL == 1 )) 00:01:13.160 + (( SPDK_TEST_NVME_FDP == 1 )) 00:01:13.160 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:01:13.160 + for nvme in "${!nvme_files[@]}" 00:01:13.160 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-multi2.img -s 4G 00:01:13.160 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:01:13.160 + for nvme in "${!nvme_files[@]}" 00:01:13.160 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-cmb.img -s 5G 00:01:13.160 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:01:13.160 + for nvme in "${!nvme_files[@]}" 00:01:13.160 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-openstack.img -s 8G 00:01:13.160 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:01:13.160 + for nvme in "${!nvme_files[@]}" 00:01:13.160 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-zns.img -s 5G 00:01:13.160 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:01:13.161 + for nvme in "${!nvme_files[@]}" 00:01:13.161 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-multi1.img -s 4G 00:01:13.161 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:01:13.161 + for nvme in "${!nvme_files[@]}" 00:01:13.161 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-multi0.img -s 4G 00:01:13.161 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:01:13.161 + for nvme in "${!nvme_files[@]}" 00:01:13.161 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme.img -s 5G 00:01:13.420 Formatting '/var/lib/libvirt/images/backends/ex1-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:01:13.420 ++ sudo grep -rl ex1-nvme.img /etc/libvirt/qemu 00:01:13.420 + echo 'End stage prepare_nvme.sh' 00:01:13.420 End stage prepare_nvme.sh 00:01:13.433 [Pipeline] sh 00:01:13.719 + DISTRO=fedora39 CPUS=10 RAM=12288 jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:01:13.719 Setup: -n 10 -s 12288 -x http://proxy-dmz.intel.com:911 -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 -b /var/lib/libvirt/images/backends/ex1-nvme.img -b /var/lib/libvirt/images/backends/ex1-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex1-nvme-multi1.img:/var/lib/libvirt/images/backends/ex1-nvme-multi2.img -H -a -v -f fedora39 00:01:13.719 00:01:13.719 DIR=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant 00:01:13.719 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest/spdk 00:01:13.719 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest 00:01:13.719 HELP=0 00:01:13.719 DRY_RUN=0 00:01:13.719 NVME_FILE=/var/lib/libvirt/images/backends/ex1-nvme.img,/var/lib/libvirt/images/backends/ex1-nvme-multi0.img, 00:01:13.719 NVME_DISKS_TYPE=nvme,nvme, 00:01:13.719 NVME_AUTO_CREATE=0 00:01:13.719 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex1-nvme-multi1.img:/var/lib/libvirt/images/backends/ex1-nvme-multi2.img, 00:01:13.719 NVME_CMB=,, 00:01:13.720 NVME_PMR=,, 00:01:13.720 NVME_ZNS=,, 00:01:13.720 NVME_MS=,, 00:01:13.720 NVME_FDP=,, 00:01:13.720 SPDK_VAGRANT_DISTRO=fedora39 00:01:13.720 SPDK_VAGRANT_VMCPU=10 00:01:13.720 SPDK_VAGRANT_VMRAM=12288 00:01:13.720 SPDK_VAGRANT_PROVIDER=libvirt 00:01:13.720 SPDK_VAGRANT_HTTP_PROXY=http://proxy-dmz.intel.com:911 00:01:13.720 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:01:13.720 SPDK_OPENSTACK_NETWORK=0 00:01:13.720 VAGRANT_PACKAGE_BOX=0 00:01:13.720 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant/Vagrantfile 00:01:13.720 FORCE_DISTRO=true 00:01:13.720 VAGRANT_BOX_VERSION= 00:01:13.720 EXTRA_VAGRANTFILES= 00:01:13.720 NIC_MODEL=virtio 00:01:13.720 00:01:13.720 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt' 00:01:13.720 /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest 00:01:15.628 Bringing machine 'default' up with 'libvirt' provider... 00:01:16.197 ==> default: Creating image (snapshot of base box volume). 00:01:16.197 ==> default: Creating domain with the following settings... 00:01:16.197 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1733549329_094cfe5b03325ce357db 00:01:16.197 ==> default: -- Domain type: kvm 00:01:16.197 ==> default: -- Cpus: 10 00:01:16.197 ==> default: -- Feature: acpi 00:01:16.197 ==> default: -- Feature: apic 00:01:16.197 ==> default: -- Feature: pae 00:01:16.197 ==> default: -- Memory: 12288M 00:01:16.197 ==> default: -- Memory Backing: hugepages: 00:01:16.197 ==> default: -- Management MAC: 00:01:16.197 ==> default: -- Loader: 00:01:16.197 ==> default: -- Nvram: 00:01:16.197 ==> default: -- Base box: spdk/fedora39 00:01:16.197 ==> default: -- Storage pool: default 00:01:16.197 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1733549329_094cfe5b03325ce357db.img (20G) 00:01:16.197 ==> default: -- Volume Cache: default 00:01:16.197 ==> default: -- Kernel: 00:01:16.197 ==> default: -- Initrd: 00:01:16.197 ==> default: -- Graphics Type: vnc 00:01:16.197 ==> default: -- Graphics Port: -1 00:01:16.197 ==> default: -- Graphics IP: 127.0.0.1 00:01:16.197 ==> default: -- Graphics Password: Not defined 00:01:16.197 ==> default: -- Video Type: cirrus 00:01:16.197 ==> default: -- Video VRAM: 9216 00:01:16.197 ==> default: -- Sound Type: 00:01:16.197 ==> default: -- Keymap: en-us 00:01:16.197 ==> default: -- TPM Path: 00:01:16.197 ==> default: -- INPUT: type=mouse, bus=ps2 00:01:16.197 ==> default: -- Command line args: 00:01:16.197 ==> default: -> value=-device, 00:01:16.197 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:01:16.197 ==> default: -> value=-drive, 00:01:16.197 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex1-nvme.img,if=none,id=nvme-0-drive0, 00:01:16.197 ==> default: -> value=-device, 00:01:16.197 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:16.197 ==> default: -> value=-device, 00:01:16.197 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:01:16.197 ==> default: -> value=-drive, 00:01:16.197 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex1-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:01:16.197 ==> default: -> value=-device, 00:01:16.197 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:16.197 ==> default: -> value=-drive, 00:01:16.197 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex1-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:01:16.197 ==> default: -> value=-device, 00:01:16.197 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:16.197 ==> default: -> value=-drive, 00:01:16.197 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex1-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:01:16.197 ==> default: -> value=-device, 00:01:16.197 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:16.456 ==> default: Creating shared folders metadata... 00:01:16.456 ==> default: Starting domain. 00:01:17.838 ==> default: Waiting for domain to get an IP address... 00:01:35.965 ==> default: Waiting for SSH to become available... 00:01:35.965 ==> default: Configuring and enabling network interfaces... 00:01:41.253 default: SSH address: 192.168.121.249:22 00:01:41.253 default: SSH username: vagrant 00:01:41.253 default: SSH auth method: private key 00:01:43.794 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/spdk/ => /home/vagrant/spdk_repo/spdk 00:01:51.925 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/dpdk/ => /home/vagrant/spdk_repo/dpdk 00:01:57.235 ==> default: Mounting SSHFS shared folder... 00:01:59.775 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:01:59.775 ==> default: Checking Mount.. 00:02:01.157 ==> default: Folder Successfully Mounted! 00:02:01.157 ==> default: Running provisioner: file... 00:02:02.094 default: ~/.gitconfig => .gitconfig 00:02:02.659 00:02:02.659 SUCCESS! 00:02:02.659 00:02:02.659 cd to /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt and type "vagrant ssh" to use. 00:02:02.659 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:02:02.659 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt" to destroy all trace of vm. 00:02:02.659 00:02:02.668 [Pipeline] } 00:02:02.682 [Pipeline] // stage 00:02:02.692 [Pipeline] dir 00:02:02.692 Running in /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt 00:02:02.693 [Pipeline] { 00:02:02.706 [Pipeline] catchError 00:02:02.708 [Pipeline] { 00:02:02.718 [Pipeline] sh 00:02:02.996 + vagrant ssh-config --host vagrant 00:02:02.996 + sed -ne /^Host/,$p 00:02:02.996 + tee ssh_conf 00:02:05.556 Host vagrant 00:02:05.556 HostName 192.168.121.249 00:02:05.556 User vagrant 00:02:05.556 Port 22 00:02:05.556 UserKnownHostsFile /dev/null 00:02:05.556 StrictHostKeyChecking no 00:02:05.556 PasswordAuthentication no 00:02:05.556 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:02:05.556 IdentitiesOnly yes 00:02:05.556 LogLevel FATAL 00:02:05.556 ForwardAgent yes 00:02:05.556 ForwardX11 yes 00:02:05.556 00:02:05.591 [Pipeline] withEnv 00:02:05.594 [Pipeline] { 00:02:05.609 [Pipeline] sh 00:02:05.895 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant #!/bin/bash 00:02:05.895 source /etc/os-release 00:02:05.895 [[ -e /image.version ]] && img=$(< /image.version) 00:02:05.895 # Minimal, systemd-like check. 00:02:05.895 if [[ -e /.dockerenv ]]; then 00:02:05.895 # Clear garbage from the node's name: 00:02:05.895 # agt-er_autotest_547-896 -> autotest_547-896 00:02:05.895 # $HOSTNAME is the actual container id 00:02:05.895 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:02:05.895 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:02:05.895 # We can assume this is a mount from a host where container is running, 00:02:05.895 # so fetch its hostname to easily identify the target swarm worker. 00:02:05.895 container="$(< /etc/hostname) ($agent)" 00:02:05.895 else 00:02:05.895 # Fallback 00:02:05.895 container=$agent 00:02:05.895 fi 00:02:05.895 fi 00:02:05.895 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:02:05.895 00:02:06.169 [Pipeline] } 00:02:06.186 [Pipeline] // withEnv 00:02:06.196 [Pipeline] setCustomBuildProperty 00:02:06.212 [Pipeline] stage 00:02:06.215 [Pipeline] { (Tests) 00:02:06.232 [Pipeline] sh 00:02:06.516 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:02:06.791 [Pipeline] sh 00:02:07.074 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:02:07.352 [Pipeline] timeout 00:02:07.353 Timeout set to expire in 1 hr 30 min 00:02:07.355 [Pipeline] { 00:02:07.371 [Pipeline] sh 00:02:07.656 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant git -C spdk_repo/spdk reset --hard 00:02:08.225 HEAD is now at a2f5e1c2d blob: don't free bs when spdk_bs_destroy/spdk_bs_unload fails 00:02:08.240 [Pipeline] sh 00:02:08.522 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant sudo chown vagrant:vagrant spdk_repo 00:02:08.797 [Pipeline] sh 00:02:09.083 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:02:09.361 [Pipeline] sh 00:02:09.652 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo 00:02:09.913 ++ readlink -f spdk_repo 00:02:09.913 + DIR_ROOT=/home/vagrant/spdk_repo 00:02:09.913 + [[ -n /home/vagrant/spdk_repo ]] 00:02:09.913 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:02:09.913 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:02:09.913 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:02:09.913 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:02:09.913 + [[ -d /home/vagrant/spdk_repo/output ]] 00:02:09.913 + [[ raid-vg-autotest == pkgdep-* ]] 00:02:09.913 + cd /home/vagrant/spdk_repo 00:02:09.913 + source /etc/os-release 00:02:09.913 ++ NAME='Fedora Linux' 00:02:09.913 ++ VERSION='39 (Cloud Edition)' 00:02:09.913 ++ ID=fedora 00:02:09.913 ++ VERSION_ID=39 00:02:09.913 ++ VERSION_CODENAME= 00:02:09.913 ++ PLATFORM_ID=platform:f39 00:02:09.913 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:02:09.913 ++ ANSI_COLOR='0;38;2;60;110;180' 00:02:09.913 ++ LOGO=fedora-logo-icon 00:02:09.913 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:02:09.913 ++ HOME_URL=https://fedoraproject.org/ 00:02:09.913 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:02:09.913 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:02:09.913 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:02:09.913 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:02:09.913 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:02:09.913 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:02:09.913 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:02:09.913 ++ SUPPORT_END=2024-11-12 00:02:09.913 ++ VARIANT='Cloud Edition' 00:02:09.913 ++ VARIANT_ID=cloud 00:02:09.913 + uname -a 00:02:09.913 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:02:09.913 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:02:10.481 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:02:10.481 Hugepages 00:02:10.481 node hugesize free / total 00:02:10.481 node0 1048576kB 0 / 0 00:02:10.481 node0 2048kB 0 / 0 00:02:10.481 00:02:10.481 Type BDF Vendor Device NUMA Driver Device Block devices 00:02:10.481 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:02:10.481 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:02:10.481 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:02:10.481 + rm -f /tmp/spdk-ld-path 00:02:10.481 + source autorun-spdk.conf 00:02:10.481 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:10.481 ++ SPDK_RUN_ASAN=1 00:02:10.481 ++ SPDK_RUN_UBSAN=1 00:02:10.481 ++ SPDK_TEST_RAID=1 00:02:10.481 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:02:10.481 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:10.481 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:10.481 ++ RUN_NIGHTLY=1 00:02:10.481 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:02:10.481 + [[ -n '' ]] 00:02:10.481 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:02:10.481 + for M in /var/spdk/build-*-manifest.txt 00:02:10.481 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:02:10.481 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:10.481 + for M in /var/spdk/build-*-manifest.txt 00:02:10.481 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:02:10.481 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:10.481 + for M in /var/spdk/build-*-manifest.txt 00:02:10.481 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:02:10.481 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:10.769 ++ uname 00:02:10.769 + [[ Linux == \L\i\n\u\x ]] 00:02:10.769 + sudo dmesg -T 00:02:10.769 + sudo dmesg --clear 00:02:10.769 + dmesg_pid=6170 00:02:10.769 + [[ Fedora Linux == FreeBSD ]] 00:02:10.769 + sudo dmesg -Tw 00:02:10.769 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:10.769 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:10.769 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:02:10.769 + [[ -x /usr/src/fio-static/fio ]] 00:02:10.769 + export FIO_BIN=/usr/src/fio-static/fio 00:02:10.769 + FIO_BIN=/usr/src/fio-static/fio 00:02:10.769 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:02:10.769 + [[ ! -v VFIO_QEMU_BIN ]] 00:02:10.769 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:02:10.769 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:10.769 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:10.769 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:02:10.769 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:10.769 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:10.769 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:10.769 05:29:44 -- common/autotest_common.sh@1710 -- $ [[ n == y ]] 00:02:10.769 05:29:44 -- spdk/autorun.sh@20 -- $ source /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:10.769 05:29:44 -- spdk_repo/autorun-spdk.conf@1 -- $ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:10.769 05:29:44 -- spdk_repo/autorun-spdk.conf@2 -- $ SPDK_RUN_ASAN=1 00:02:10.769 05:29:44 -- spdk_repo/autorun-spdk.conf@3 -- $ SPDK_RUN_UBSAN=1 00:02:10.769 05:29:44 -- spdk_repo/autorun-spdk.conf@4 -- $ SPDK_TEST_RAID=1 00:02:10.769 05:29:44 -- spdk_repo/autorun-spdk.conf@5 -- $ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:02:10.769 05:29:44 -- spdk_repo/autorun-spdk.conf@6 -- $ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:10.769 05:29:44 -- spdk_repo/autorun-spdk.conf@7 -- $ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:10.769 05:29:44 -- spdk_repo/autorun-spdk.conf@8 -- $ RUN_NIGHTLY=1 00:02:10.769 05:29:44 -- spdk/autorun.sh@22 -- $ trap 'timing_finish || exit 1' EXIT 00:02:10.769 05:29:44 -- spdk/autorun.sh@25 -- $ /home/vagrant/spdk_repo/spdk/autobuild.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:10.769 05:29:44 -- common/autotest_common.sh@1710 -- $ [[ n == y ]] 00:02:10.769 05:29:44 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:02:10.769 05:29:44 -- scripts/common.sh@15 -- $ shopt -s extglob 00:02:11.028 05:29:44 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:02:11.028 05:29:44 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:02:11.028 05:29:44 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:02:11.028 05:29:44 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:11.028 05:29:44 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:11.028 05:29:44 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:11.028 05:29:44 -- paths/export.sh@5 -- $ export PATH 00:02:11.028 05:29:44 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:11.028 05:29:44 -- common/autobuild_common.sh@492 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:02:11.028 05:29:44 -- common/autobuild_common.sh@493 -- $ date +%s 00:02:11.029 05:29:44 -- common/autobuild_common.sh@493 -- $ mktemp -dt spdk_1733549384.XXXXXX 00:02:11.029 05:29:44 -- common/autobuild_common.sh@493 -- $ SPDK_WORKSPACE=/tmp/spdk_1733549384.fpHOoA 00:02:11.029 05:29:44 -- common/autobuild_common.sh@495 -- $ [[ -n '' ]] 00:02:11.029 05:29:44 -- common/autobuild_common.sh@499 -- $ '[' -n v22.11.4 ']' 00:02:11.029 05:29:44 -- common/autobuild_common.sh@500 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:11.029 05:29:44 -- common/autobuild_common.sh@500 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:02:11.029 05:29:44 -- common/autobuild_common.sh@506 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:02:11.029 05:29:44 -- common/autobuild_common.sh@508 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:02:11.029 05:29:44 -- common/autobuild_common.sh@509 -- $ get_config_params 00:02:11.029 05:29:44 -- common/autotest_common.sh@409 -- $ xtrace_disable 00:02:11.029 05:29:44 -- common/autotest_common.sh@10 -- $ set +x 00:02:11.029 05:29:44 -- common/autobuild_common.sh@509 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:02:11.029 05:29:44 -- common/autobuild_common.sh@511 -- $ start_monitor_resources 00:02:11.029 05:29:44 -- pm/common@17 -- $ local monitor 00:02:11.029 05:29:44 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:11.029 05:29:44 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:11.029 05:29:44 -- pm/common@25 -- $ sleep 1 00:02:11.029 05:29:44 -- pm/common@21 -- $ date +%s 00:02:11.029 05:29:44 -- pm/common@21 -- $ date +%s 00:02:11.029 05:29:44 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1733549384 00:02:11.029 05:29:44 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1733549384 00:02:11.029 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1733549384_collect-cpu-load.pm.log 00:02:11.029 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1733549384_collect-vmstat.pm.log 00:02:11.971 05:29:45 -- common/autobuild_common.sh@512 -- $ trap stop_monitor_resources EXIT 00:02:11.971 05:29:45 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:02:11.971 05:29:45 -- spdk/autobuild.sh@12 -- $ umask 022 00:02:11.971 05:29:45 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:02:11.971 05:29:45 -- spdk/autobuild.sh@16 -- $ date -u 00:02:11.971 Sat Dec 7 05:29:45 AM UTC 2024 00:02:11.971 05:29:45 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:02:11.971 v25.01-pre-311-ga2f5e1c2d 00:02:11.971 05:29:45 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:02:11.971 05:29:45 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:02:11.971 05:29:45 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:02:11.971 05:29:45 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:02:11.971 05:29:45 -- common/autotest_common.sh@10 -- $ set +x 00:02:11.971 ************************************ 00:02:11.971 START TEST asan 00:02:11.971 ************************************ 00:02:11.971 using asan 00:02:11.971 05:29:45 asan -- common/autotest_common.sh@1129 -- $ echo 'using asan' 00:02:11.971 00:02:11.971 real 0m0.001s 00:02:11.971 user 0m0.000s 00:02:11.971 sys 0m0.000s 00:02:11.971 05:29:45 asan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:02:11.971 05:29:45 asan -- common/autotest_common.sh@10 -- $ set +x 00:02:11.971 ************************************ 00:02:11.971 END TEST asan 00:02:11.971 ************************************ 00:02:11.971 05:29:45 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:02:11.971 05:29:45 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:02:11.971 05:29:45 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:02:11.971 05:29:45 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:02:11.971 05:29:45 -- common/autotest_common.sh@10 -- $ set +x 00:02:11.971 ************************************ 00:02:11.971 START TEST ubsan 00:02:11.971 ************************************ 00:02:11.971 using ubsan 00:02:11.971 05:29:45 ubsan -- common/autotest_common.sh@1129 -- $ echo 'using ubsan' 00:02:11.971 00:02:11.971 real 0m0.000s 00:02:11.971 user 0m0.000s 00:02:11.971 sys 0m0.000s 00:02:11.971 05:29:45 ubsan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:02:11.971 05:29:45 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:02:11.971 ************************************ 00:02:11.971 END TEST ubsan 00:02:11.971 ************************************ 00:02:12.233 05:29:45 -- spdk/autobuild.sh@27 -- $ '[' -n v22.11.4 ']' 00:02:12.233 05:29:45 -- spdk/autobuild.sh@28 -- $ build_native_dpdk 00:02:12.233 05:29:45 -- common/autobuild_common.sh@449 -- $ run_test build_native_dpdk _build_native_dpdk 00:02:12.233 05:29:45 -- common/autotest_common.sh@1105 -- $ '[' 2 -le 1 ']' 00:02:12.233 05:29:45 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:02:12.233 05:29:45 -- common/autotest_common.sh@10 -- $ set +x 00:02:12.233 ************************************ 00:02:12.233 START TEST build_native_dpdk 00:02:12.233 ************************************ 00:02:12.233 05:29:45 build_native_dpdk -- common/autotest_common.sh@1129 -- $ _build_native_dpdk 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@48 -- $ local external_dpdk_dir 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@49 -- $ local external_dpdk_base_dir 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@50 -- $ local compiler_version 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@51 -- $ local compiler 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@52 -- $ local dpdk_kmods 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@53 -- $ local repo=dpdk 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@55 -- $ compiler=gcc 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@61 -- $ export CC=gcc 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@61 -- $ CC=gcc 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *clang* ]] 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *gcc* ]] 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@68 -- $ gcc -dumpversion 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@68 -- $ compiler_version=13 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@69 -- $ compiler_version=13 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@70 -- $ external_dpdk_dir=/home/vagrant/spdk_repo/dpdk/build 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@71 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@71 -- $ external_dpdk_base_dir=/home/vagrant/spdk_repo/dpdk 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@73 -- $ [[ ! -d /home/vagrant/spdk_repo/dpdk ]] 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@82 -- $ orgdir=/home/vagrant/spdk_repo/spdk 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@83 -- $ git -C /home/vagrant/spdk_repo/dpdk log --oneline -n 5 00:02:12.233 caf0f5d395 version: 22.11.4 00:02:12.233 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:02:12.233 dc9c799c7d vhost: fix missing spinlock unlock 00:02:12.233 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:02:12.233 6ef77f2a5e net/gve: fix RX buffer size alignment 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@85 -- $ dpdk_cflags='-fPIC -g -fcommon' 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@86 -- $ dpdk_ldflags= 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@87 -- $ dpdk_ver=22.11.4 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ gcc == *gcc* ]] 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ 13 -ge 5 ]] 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@90 -- $ dpdk_cflags+=' -Werror' 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ gcc == *gcc* ]] 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ 13 -ge 10 ]] 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@94 -- $ dpdk_cflags+=' -Wno-stringop-overflow' 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@102 -- $ DPDK_DRIVERS=("bus" "bus/pci" "bus/vdev" "mempool/ring" "net/i40e" "net/i40e/base" "power/acpi" "power/amd_pstate" "power/cppc" "power/intel_pstate" "power/intel_uncore" "power/kvm_vm") 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@103 -- $ local mlx5_libs_added=n 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@104 -- $ [[ 0 -eq 1 ]] 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@104 -- $ [[ 0 -eq 1 ]] 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@146 -- $ [[ 0 -eq 1 ]] 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@174 -- $ cd /home/vagrant/spdk_repo/dpdk 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@175 -- $ uname -s 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@175 -- $ '[' Linux = Linux ']' 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@176 -- $ lt 22.11.4 21.11.0 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 21.11.0 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 21 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@353 -- $ local d=21 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 21 =~ ^[0-9]+$ ]] 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@355 -- $ echo 21 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=21 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@367 -- $ return 1 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@180 -- $ patch -p1 00:02:12.233 patching file config/rte_config.h 00:02:12.233 Hunk #1 succeeded at 60 (offset 1 line). 00:02:12.233 05:29:45 build_native_dpdk -- common/autobuild_common.sh@183 -- $ lt 22.11.4 24.07.0 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 24.07.0 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:12.233 05:29:45 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@368 -- $ return 0 00:02:12.234 05:29:45 build_native_dpdk -- common/autobuild_common.sh@184 -- $ patch -p1 00:02:12.234 patching file lib/pcapng/rte_pcapng.c 00:02:12.234 Hunk #1 succeeded at 110 (offset -18 lines). 00:02:12.234 05:29:45 build_native_dpdk -- common/autobuild_common.sh@186 -- $ ge 22.11.4 24.07.0 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@376 -- $ cmp_versions 22.11.4 '>=' 24.07.0 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=>=' 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@348 -- $ : 1 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:12.234 05:29:45 build_native_dpdk -- scripts/common.sh@368 -- $ return 1 00:02:12.234 05:29:45 build_native_dpdk -- common/autobuild_common.sh@190 -- $ dpdk_kmods=false 00:02:12.234 05:29:45 build_native_dpdk -- common/autobuild_common.sh@191 -- $ uname -s 00:02:12.234 05:29:45 build_native_dpdk -- common/autobuild_common.sh@191 -- $ '[' Linux = FreeBSD ']' 00:02:12.234 05:29:45 build_native_dpdk -- common/autobuild_common.sh@195 -- $ printf %s, bus bus/pci bus/vdev mempool/ring net/i40e net/i40e/base power/acpi power/amd_pstate power/cppc power/intel_pstate power/intel_uncore power/kvm_vm 00:02:12.234 05:29:45 build_native_dpdk -- common/autobuild_common.sh@195 -- $ meson build-tmp --prefix=/home/vagrant/spdk_repo/dpdk/build --libdir lib -Denable_docs=false -Denable_kmods=false -Dtests=false -Dc_link_args= '-Dc_args=-fPIC -g -fcommon -Werror -Wno-stringop-overflow' -Dmachine=native -Denable_drivers=bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm, 00:02:17.513 The Meson build system 00:02:17.513 Version: 1.5.0 00:02:17.513 Source dir: /home/vagrant/spdk_repo/dpdk 00:02:17.513 Build dir: /home/vagrant/spdk_repo/dpdk/build-tmp 00:02:17.513 Build type: native build 00:02:17.513 Program cat found: YES (/usr/bin/cat) 00:02:17.513 Project name: DPDK 00:02:17.513 Project version: 22.11.4 00:02:17.513 C compiler for the host machine: gcc (gcc 13.3.1 "gcc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:02:17.513 C linker for the host machine: gcc ld.bfd 2.40-14 00:02:17.513 Host machine cpu family: x86_64 00:02:17.513 Host machine cpu: x86_64 00:02:17.513 Message: ## Building in Developer Mode ## 00:02:17.513 Program pkg-config found: YES (/usr/bin/pkg-config) 00:02:17.513 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/check-symbols.sh) 00:02:17.513 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/options-ibverbs-static.sh) 00:02:17.513 Program objdump found: YES (/usr/bin/objdump) 00:02:17.513 Program python3 found: YES (/usr/bin/python3) 00:02:17.513 Program cat found: YES (/usr/bin/cat) 00:02:17.513 config/meson.build:83: WARNING: The "machine" option is deprecated. Please use "cpu_instruction_set" instead. 00:02:17.513 Checking for size of "void *" : 8 00:02:17.513 Checking for size of "void *" : 8 (cached) 00:02:17.513 Library m found: YES 00:02:17.513 Library numa found: YES 00:02:17.513 Has header "numaif.h" : YES 00:02:17.513 Library fdt found: NO 00:02:17.513 Library execinfo found: NO 00:02:17.513 Has header "execinfo.h" : YES 00:02:17.513 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:02:17.513 Run-time dependency libarchive found: NO (tried pkgconfig) 00:02:17.513 Run-time dependency libbsd found: NO (tried pkgconfig) 00:02:17.513 Run-time dependency jansson found: NO (tried pkgconfig) 00:02:17.513 Run-time dependency openssl found: YES 3.1.1 00:02:17.513 Run-time dependency libpcap found: YES 1.10.4 00:02:17.513 Has header "pcap.h" with dependency libpcap: YES 00:02:17.513 Compiler for C supports arguments -Wcast-qual: YES 00:02:17.513 Compiler for C supports arguments -Wdeprecated: YES 00:02:17.513 Compiler for C supports arguments -Wformat: YES 00:02:17.513 Compiler for C supports arguments -Wformat-nonliteral: NO 00:02:17.513 Compiler for C supports arguments -Wformat-security: NO 00:02:17.513 Compiler for C supports arguments -Wmissing-declarations: YES 00:02:17.513 Compiler for C supports arguments -Wmissing-prototypes: YES 00:02:17.513 Compiler for C supports arguments -Wnested-externs: YES 00:02:17.513 Compiler for C supports arguments -Wold-style-definition: YES 00:02:17.513 Compiler for C supports arguments -Wpointer-arith: YES 00:02:17.513 Compiler for C supports arguments -Wsign-compare: YES 00:02:17.513 Compiler for C supports arguments -Wstrict-prototypes: YES 00:02:17.513 Compiler for C supports arguments -Wundef: YES 00:02:17.513 Compiler for C supports arguments -Wwrite-strings: YES 00:02:17.513 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:02:17.513 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:02:17.513 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:02:17.513 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:02:17.513 Compiler for C supports arguments -mavx512f: YES 00:02:17.513 Checking if "AVX512 checking" compiles: YES 00:02:17.513 Fetching value of define "__SSE4_2__" : 1 00:02:17.513 Fetching value of define "__AES__" : 1 00:02:17.513 Fetching value of define "__AVX__" : 1 00:02:17.513 Fetching value of define "__AVX2__" : 1 00:02:17.513 Fetching value of define "__AVX512BW__" : 1 00:02:17.513 Fetching value of define "__AVX512CD__" : 1 00:02:17.513 Fetching value of define "__AVX512DQ__" : 1 00:02:17.513 Fetching value of define "__AVX512F__" : 1 00:02:17.513 Fetching value of define "__AVX512VL__" : 1 00:02:17.513 Fetching value of define "__PCLMUL__" : 1 00:02:17.513 Fetching value of define "__RDRND__" : 1 00:02:17.513 Fetching value of define "__RDSEED__" : 1 00:02:17.513 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:02:17.513 Compiler for C supports arguments -Wno-format-truncation: YES 00:02:17.513 Message: lib/kvargs: Defining dependency "kvargs" 00:02:17.513 Message: lib/telemetry: Defining dependency "telemetry" 00:02:17.513 Checking for function "getentropy" : YES 00:02:17.513 Message: lib/eal: Defining dependency "eal" 00:02:17.513 Message: lib/ring: Defining dependency "ring" 00:02:17.513 Message: lib/rcu: Defining dependency "rcu" 00:02:17.513 Message: lib/mempool: Defining dependency "mempool" 00:02:17.513 Message: lib/mbuf: Defining dependency "mbuf" 00:02:17.513 Fetching value of define "__PCLMUL__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:17.513 Fetching value of define "__VPCLMULQDQ__" : (undefined) (cached) 00:02:17.513 Compiler for C supports arguments -mpclmul: YES 00:02:17.513 Compiler for C supports arguments -maes: YES 00:02:17.513 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:17.513 Compiler for C supports arguments -mavx512bw: YES 00:02:17.513 Compiler for C supports arguments -mavx512dq: YES 00:02:17.513 Compiler for C supports arguments -mavx512vl: YES 00:02:17.513 Compiler for C supports arguments -mvpclmulqdq: YES 00:02:17.513 Compiler for C supports arguments -mavx2: YES 00:02:17.513 Compiler for C supports arguments -mavx: YES 00:02:17.513 Message: lib/net: Defining dependency "net" 00:02:17.513 Message: lib/meter: Defining dependency "meter" 00:02:17.513 Message: lib/ethdev: Defining dependency "ethdev" 00:02:17.513 Message: lib/pci: Defining dependency "pci" 00:02:17.513 Message: lib/cmdline: Defining dependency "cmdline" 00:02:17.513 Message: lib/metrics: Defining dependency "metrics" 00:02:17.513 Message: lib/hash: Defining dependency "hash" 00:02:17.513 Message: lib/timer: Defining dependency "timer" 00:02:17.513 Fetching value of define "__AVX2__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512CD__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:17.513 Message: lib/acl: Defining dependency "acl" 00:02:17.513 Message: lib/bbdev: Defining dependency "bbdev" 00:02:17.513 Message: lib/bitratestats: Defining dependency "bitratestats" 00:02:17.513 Run-time dependency libelf found: YES 0.191 00:02:17.513 Message: lib/bpf: Defining dependency "bpf" 00:02:17.513 Message: lib/cfgfile: Defining dependency "cfgfile" 00:02:17.513 Message: lib/compressdev: Defining dependency "compressdev" 00:02:17.513 Message: lib/cryptodev: Defining dependency "cryptodev" 00:02:17.513 Message: lib/distributor: Defining dependency "distributor" 00:02:17.513 Message: lib/efd: Defining dependency "efd" 00:02:17.513 Message: lib/eventdev: Defining dependency "eventdev" 00:02:17.513 Message: lib/gpudev: Defining dependency "gpudev" 00:02:17.513 Message: lib/gro: Defining dependency "gro" 00:02:17.513 Message: lib/gso: Defining dependency "gso" 00:02:17.513 Message: lib/ip_frag: Defining dependency "ip_frag" 00:02:17.513 Message: lib/jobstats: Defining dependency "jobstats" 00:02:17.513 Message: lib/latencystats: Defining dependency "latencystats" 00:02:17.513 Message: lib/lpm: Defining dependency "lpm" 00:02:17.513 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512IFMA__" : (undefined) 00:02:17.513 Compiler for C supports arguments -mavx512f -mavx512dq -mavx512ifma: YES 00:02:17.513 Message: lib/member: Defining dependency "member" 00:02:17.513 Message: lib/pcapng: Defining dependency "pcapng" 00:02:17.513 Compiler for C supports arguments -Wno-cast-qual: YES 00:02:17.513 Message: lib/power: Defining dependency "power" 00:02:17.513 Message: lib/rawdev: Defining dependency "rawdev" 00:02:17.513 Message: lib/regexdev: Defining dependency "regexdev" 00:02:17.513 Message: lib/dmadev: Defining dependency "dmadev" 00:02:17.513 Message: lib/rib: Defining dependency "rib" 00:02:17.513 Message: lib/reorder: Defining dependency "reorder" 00:02:17.513 Message: lib/sched: Defining dependency "sched" 00:02:17.513 Message: lib/security: Defining dependency "security" 00:02:17.513 Message: lib/stack: Defining dependency "stack" 00:02:17.513 Has header "linux/userfaultfd.h" : YES 00:02:17.513 Message: lib/vhost: Defining dependency "vhost" 00:02:17.513 Message: lib/ipsec: Defining dependency "ipsec" 00:02:17.513 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:17.513 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:17.513 Message: lib/fib: Defining dependency "fib" 00:02:17.513 Message: lib/port: Defining dependency "port" 00:02:17.513 Message: lib/pdump: Defining dependency "pdump" 00:02:17.513 Message: lib/table: Defining dependency "table" 00:02:17.513 Message: lib/pipeline: Defining dependency "pipeline" 00:02:17.513 Message: lib/graph: Defining dependency "graph" 00:02:17.513 Message: lib/node: Defining dependency "node" 00:02:17.513 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:02:17.513 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:02:17.513 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:02:17.513 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:02:17.513 Compiler for C supports arguments -Wno-sign-compare: YES 00:02:17.513 Compiler for C supports arguments -Wno-unused-value: YES 00:02:17.513 Compiler for C supports arguments -Wno-format: YES 00:02:17.513 Compiler for C supports arguments -Wno-format-security: YES 00:02:17.513 Compiler for C supports arguments -Wno-format-nonliteral: YES 00:02:17.513 Compiler for C supports arguments -Wno-strict-aliasing: YES 00:02:19.007 Compiler for C supports arguments -Wno-unused-but-set-variable: YES 00:02:19.007 Compiler for C supports arguments -Wno-unused-parameter: YES 00:02:19.007 Fetching value of define "__AVX2__" : 1 (cached) 00:02:19.007 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:19.007 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:19.007 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:19.007 Compiler for C supports arguments -mavx512bw: YES (cached) 00:02:19.007 Compiler for C supports arguments -march=skylake-avx512: YES 00:02:19.007 Message: drivers/net/i40e: Defining dependency "net_i40e" 00:02:19.007 Program doxygen found: YES (/usr/local/bin/doxygen) 00:02:19.007 Configuring doxy-api.conf using configuration 00:02:19.007 Program sphinx-build found: NO 00:02:19.008 Configuring rte_build_config.h using configuration 00:02:19.008 Message: 00:02:19.008 ================= 00:02:19.008 Applications Enabled 00:02:19.008 ================= 00:02:19.008 00:02:19.008 apps: 00:02:19.008 dumpcap, pdump, proc-info, test-acl, test-bbdev, test-cmdline, test-compress-perf, test-crypto-perf, 00:02:19.008 test-eventdev, test-fib, test-flow-perf, test-gpudev, test-pipeline, test-pmd, test-regex, test-sad, 00:02:19.008 test-security-perf, 00:02:19.008 00:02:19.008 Message: 00:02:19.008 ================= 00:02:19.008 Libraries Enabled 00:02:19.008 ================= 00:02:19.008 00:02:19.008 libs: 00:02:19.008 kvargs, telemetry, eal, ring, rcu, mempool, mbuf, net, 00:02:19.008 meter, ethdev, pci, cmdline, metrics, hash, timer, acl, 00:02:19.008 bbdev, bitratestats, bpf, cfgfile, compressdev, cryptodev, distributor, efd, 00:02:19.008 eventdev, gpudev, gro, gso, ip_frag, jobstats, latencystats, lpm, 00:02:19.008 member, pcapng, power, rawdev, regexdev, dmadev, rib, reorder, 00:02:19.008 sched, security, stack, vhost, ipsec, fib, port, pdump, 00:02:19.008 table, pipeline, graph, node, 00:02:19.008 00:02:19.008 Message: 00:02:19.008 =============== 00:02:19.008 Drivers Enabled 00:02:19.008 =============== 00:02:19.008 00:02:19.008 common: 00:02:19.008 00:02:19.008 bus: 00:02:19.008 pci, vdev, 00:02:19.008 mempool: 00:02:19.008 ring, 00:02:19.008 dma: 00:02:19.008 00:02:19.008 net: 00:02:19.008 i40e, 00:02:19.008 raw: 00:02:19.008 00:02:19.008 crypto: 00:02:19.008 00:02:19.008 compress: 00:02:19.008 00:02:19.008 regex: 00:02:19.008 00:02:19.008 vdpa: 00:02:19.008 00:02:19.008 event: 00:02:19.008 00:02:19.008 baseband: 00:02:19.008 00:02:19.008 gpu: 00:02:19.008 00:02:19.008 00:02:19.008 Message: 00:02:19.008 ================= 00:02:19.008 Content Skipped 00:02:19.008 ================= 00:02:19.008 00:02:19.008 apps: 00:02:19.008 00:02:19.008 libs: 00:02:19.008 kni: explicitly disabled via build config (deprecated lib) 00:02:19.008 flow_classify: explicitly disabled via build config (deprecated lib) 00:02:19.008 00:02:19.008 drivers: 00:02:19.008 common/cpt: not in enabled drivers build config 00:02:19.008 common/dpaax: not in enabled drivers build config 00:02:19.008 common/iavf: not in enabled drivers build config 00:02:19.008 common/idpf: not in enabled drivers build config 00:02:19.008 common/mvep: not in enabled drivers build config 00:02:19.008 common/octeontx: not in enabled drivers build config 00:02:19.008 bus/auxiliary: not in enabled drivers build config 00:02:19.008 bus/dpaa: not in enabled drivers build config 00:02:19.008 bus/fslmc: not in enabled drivers build config 00:02:19.008 bus/ifpga: not in enabled drivers build config 00:02:19.008 bus/vmbus: not in enabled drivers build config 00:02:19.008 common/cnxk: not in enabled drivers build config 00:02:19.008 common/mlx5: not in enabled drivers build config 00:02:19.008 common/qat: not in enabled drivers build config 00:02:19.008 common/sfc_efx: not in enabled drivers build config 00:02:19.008 mempool/bucket: not in enabled drivers build config 00:02:19.008 mempool/cnxk: not in enabled drivers build config 00:02:19.008 mempool/dpaa: not in enabled drivers build config 00:02:19.008 mempool/dpaa2: not in enabled drivers build config 00:02:19.008 mempool/octeontx: not in enabled drivers build config 00:02:19.008 mempool/stack: not in enabled drivers build config 00:02:19.008 dma/cnxk: not in enabled drivers build config 00:02:19.008 dma/dpaa: not in enabled drivers build config 00:02:19.008 dma/dpaa2: not in enabled drivers build config 00:02:19.008 dma/hisilicon: not in enabled drivers build config 00:02:19.008 dma/idxd: not in enabled drivers build config 00:02:19.008 dma/ioat: not in enabled drivers build config 00:02:19.008 dma/skeleton: not in enabled drivers build config 00:02:19.008 net/af_packet: not in enabled drivers build config 00:02:19.008 net/af_xdp: not in enabled drivers build config 00:02:19.008 net/ark: not in enabled drivers build config 00:02:19.008 net/atlantic: not in enabled drivers build config 00:02:19.008 net/avp: not in enabled drivers build config 00:02:19.008 net/axgbe: not in enabled drivers build config 00:02:19.008 net/bnx2x: not in enabled drivers build config 00:02:19.008 net/bnxt: not in enabled drivers build config 00:02:19.008 net/bonding: not in enabled drivers build config 00:02:19.008 net/cnxk: not in enabled drivers build config 00:02:19.008 net/cxgbe: not in enabled drivers build config 00:02:19.008 net/dpaa: not in enabled drivers build config 00:02:19.008 net/dpaa2: not in enabled drivers build config 00:02:19.008 net/e1000: not in enabled drivers build config 00:02:19.008 net/ena: not in enabled drivers build config 00:02:19.008 net/enetc: not in enabled drivers build config 00:02:19.008 net/enetfec: not in enabled drivers build config 00:02:19.008 net/enic: not in enabled drivers build config 00:02:19.008 net/failsafe: not in enabled drivers build config 00:02:19.008 net/fm10k: not in enabled drivers build config 00:02:19.008 net/gve: not in enabled drivers build config 00:02:19.008 net/hinic: not in enabled drivers build config 00:02:19.008 net/hns3: not in enabled drivers build config 00:02:19.008 net/iavf: not in enabled drivers build config 00:02:19.008 net/ice: not in enabled drivers build config 00:02:19.008 net/idpf: not in enabled drivers build config 00:02:19.008 net/igc: not in enabled drivers build config 00:02:19.008 net/ionic: not in enabled drivers build config 00:02:19.008 net/ipn3ke: not in enabled drivers build config 00:02:19.008 net/ixgbe: not in enabled drivers build config 00:02:19.008 net/kni: not in enabled drivers build config 00:02:19.008 net/liquidio: not in enabled drivers build config 00:02:19.008 net/mana: not in enabled drivers build config 00:02:19.008 net/memif: not in enabled drivers build config 00:02:19.008 net/mlx4: not in enabled drivers build config 00:02:19.008 net/mlx5: not in enabled drivers build config 00:02:19.008 net/mvneta: not in enabled drivers build config 00:02:19.008 net/mvpp2: not in enabled drivers build config 00:02:19.008 net/netvsc: not in enabled drivers build config 00:02:19.008 net/nfb: not in enabled drivers build config 00:02:19.008 net/nfp: not in enabled drivers build config 00:02:19.008 net/ngbe: not in enabled drivers build config 00:02:19.008 net/null: not in enabled drivers build config 00:02:19.008 net/octeontx: not in enabled drivers build config 00:02:19.008 net/octeon_ep: not in enabled drivers build config 00:02:19.008 net/pcap: not in enabled drivers build config 00:02:19.008 net/pfe: not in enabled drivers build config 00:02:19.008 net/qede: not in enabled drivers build config 00:02:19.008 net/ring: not in enabled drivers build config 00:02:19.008 net/sfc: not in enabled drivers build config 00:02:19.008 net/softnic: not in enabled drivers build config 00:02:19.008 net/tap: not in enabled drivers build config 00:02:19.008 net/thunderx: not in enabled drivers build config 00:02:19.008 net/txgbe: not in enabled drivers build config 00:02:19.008 net/vdev_netvsc: not in enabled drivers build config 00:02:19.008 net/vhost: not in enabled drivers build config 00:02:19.008 net/virtio: not in enabled drivers build config 00:02:19.008 net/vmxnet3: not in enabled drivers build config 00:02:19.008 raw/cnxk_bphy: not in enabled drivers build config 00:02:19.008 raw/cnxk_gpio: not in enabled drivers build config 00:02:19.008 raw/dpaa2_cmdif: not in enabled drivers build config 00:02:19.008 raw/ifpga: not in enabled drivers build config 00:02:19.008 raw/ntb: not in enabled drivers build config 00:02:19.008 raw/skeleton: not in enabled drivers build config 00:02:19.008 crypto/armv8: not in enabled drivers build config 00:02:19.008 crypto/bcmfs: not in enabled drivers build config 00:02:19.008 crypto/caam_jr: not in enabled drivers build config 00:02:19.008 crypto/ccp: not in enabled drivers build config 00:02:19.008 crypto/cnxk: not in enabled drivers build config 00:02:19.008 crypto/dpaa_sec: not in enabled drivers build config 00:02:19.008 crypto/dpaa2_sec: not in enabled drivers build config 00:02:19.008 crypto/ipsec_mb: not in enabled drivers build config 00:02:19.008 crypto/mlx5: not in enabled drivers build config 00:02:19.008 crypto/mvsam: not in enabled drivers build config 00:02:19.008 crypto/nitrox: not in enabled drivers build config 00:02:19.008 crypto/null: not in enabled drivers build config 00:02:19.008 crypto/octeontx: not in enabled drivers build config 00:02:19.008 crypto/openssl: not in enabled drivers build config 00:02:19.008 crypto/scheduler: not in enabled drivers build config 00:02:19.008 crypto/uadk: not in enabled drivers build config 00:02:19.008 crypto/virtio: not in enabled drivers build config 00:02:19.008 compress/isal: not in enabled drivers build config 00:02:19.008 compress/mlx5: not in enabled drivers build config 00:02:19.008 compress/octeontx: not in enabled drivers build config 00:02:19.008 compress/zlib: not in enabled drivers build config 00:02:19.008 regex/mlx5: not in enabled drivers build config 00:02:19.008 regex/cn9k: not in enabled drivers build config 00:02:19.008 vdpa/ifc: not in enabled drivers build config 00:02:19.008 vdpa/mlx5: not in enabled drivers build config 00:02:19.008 vdpa/sfc: not in enabled drivers build config 00:02:19.008 event/cnxk: not in enabled drivers build config 00:02:19.008 event/dlb2: not in enabled drivers build config 00:02:19.008 event/dpaa: not in enabled drivers build config 00:02:19.008 event/dpaa2: not in enabled drivers build config 00:02:19.008 event/dsw: not in enabled drivers build config 00:02:19.008 event/opdl: not in enabled drivers build config 00:02:19.008 event/skeleton: not in enabled drivers build config 00:02:19.008 event/sw: not in enabled drivers build config 00:02:19.008 event/octeontx: not in enabled drivers build config 00:02:19.008 baseband/acc: not in enabled drivers build config 00:02:19.008 baseband/fpga_5gnr_fec: not in enabled drivers build config 00:02:19.008 baseband/fpga_lte_fec: not in enabled drivers build config 00:02:19.008 baseband/la12xx: not in enabled drivers build config 00:02:19.008 baseband/null: not in enabled drivers build config 00:02:19.008 baseband/turbo_sw: not in enabled drivers build config 00:02:19.008 gpu/cuda: not in enabled drivers build config 00:02:19.008 00:02:19.009 00:02:19.009 Build targets in project: 311 00:02:19.009 00:02:19.009 DPDK 22.11.4 00:02:19.009 00:02:19.009 User defined options 00:02:19.009 libdir : lib 00:02:19.009 prefix : /home/vagrant/spdk_repo/dpdk/build 00:02:19.009 c_args : -fPIC -g -fcommon -Werror -Wno-stringop-overflow 00:02:19.009 c_link_args : 00:02:19.009 enable_docs : false 00:02:19.009 enable_drivers: bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm, 00:02:19.009 enable_kmods : false 00:02:19.009 machine : native 00:02:19.009 tests : false 00:02:19.009 00:02:19.009 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:02:19.009 WARNING: Running the setup command as `meson [options]` instead of `meson setup [options]` is ambiguous and deprecated. 00:02:19.009 05:29:52 build_native_dpdk -- common/autobuild_common.sh@199 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 00:02:19.009 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:02:19.268 [1/740] Generating lib/rte_kvargs_def with a custom command 00:02:19.268 [2/740] Generating lib/rte_telemetry_mingw with a custom command 00:02:19.268 [3/740] Generating lib/rte_telemetry_def with a custom command 00:02:19.268 [4/740] Generating lib/rte_kvargs_mingw with a custom command 00:02:19.268 [5/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:02:19.268 [6/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:02:19.268 [7/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:02:19.268 [8/740] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:02:19.268 [9/740] Linking static target lib/librte_kvargs.a 00:02:19.268 [10/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:02:19.268 [11/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:02:19.268 [12/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:02:19.268 [13/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:02:19.268 [14/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:02:19.268 [15/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:02:19.268 [16/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:02:19.268 [17/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:02:19.268 [18/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:02:19.528 [19/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:02:19.528 [20/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_log.c.o 00:02:19.528 [21/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:02:19.528 [22/740] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:02:19.528 [23/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:02:19.528 [24/740] Linking target lib/librte_kvargs.so.23.0 00:02:19.528 [25/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:02:19.528 [26/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:02:19.528 [27/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:02:19.528 [28/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:02:19.528 [29/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:02:19.528 [30/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:02:19.528 [31/740] Linking static target lib/librte_telemetry.a 00:02:19.528 [32/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:02:19.528 [33/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:02:19.528 [34/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:02:19.788 [35/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:02:19.788 [36/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:02:19.788 [37/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:02:19.788 [38/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:02:19.788 [39/740] Generating symbol file lib/librte_kvargs.so.23.0.p/librte_kvargs.so.23.0.symbols 00:02:19.788 [40/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:02:19.788 [41/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:02:19.788 [42/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:02:19.788 [43/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:02:20.048 [44/740] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:02:20.048 [45/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:02:20.048 [46/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:02:20.048 [47/740] Linking target lib/librte_telemetry.so.23.0 00:02:20.048 [48/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:02:20.048 [49/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:02:20.048 [50/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:02:20.048 [51/740] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:02:20.048 [52/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:02:20.048 [53/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:02:20.048 [54/740] Generating symbol file lib/librte_telemetry.so.23.0.p/librte_telemetry.so.23.0.symbols 00:02:20.048 [55/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:02:20.048 [56/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:02:20.048 [57/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:02:20.048 [58/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:02:20.049 [59/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:02:20.049 [60/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:02:20.049 [61/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:02:20.049 [62/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:02:20.049 [63/740] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:02:20.049 [64/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:02:20.049 [65/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:02:20.308 [66/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_log.c.o 00:02:20.308 [67/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:02:20.309 [68/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:02:20.309 [69/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:02:20.309 [70/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:02:20.309 [71/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:02:20.309 [72/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:02:20.309 [73/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:02:20.309 [74/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:02:20.309 [75/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:02:20.309 [76/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:02:20.309 [77/740] Generating lib/rte_eal_def with a custom command 00:02:20.309 [78/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:02:20.309 [79/740] Generating lib/rte_eal_mingw with a custom command 00:02:20.309 [80/740] Generating lib/rte_ring_def with a custom command 00:02:20.309 [81/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:02:20.309 [82/740] Generating lib/rte_ring_mingw with a custom command 00:02:20.309 [83/740] Generating lib/rte_rcu_def with a custom command 00:02:20.309 [84/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:02:20.309 [85/740] Generating lib/rte_rcu_mingw with a custom command 00:02:20.309 [86/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:02:20.569 [87/740] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:02:20.569 [88/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:02:20.569 [89/740] Linking static target lib/librte_ring.a 00:02:20.569 [90/740] Generating lib/rte_mempool_def with a custom command 00:02:20.569 [91/740] Generating lib/rte_mempool_mingw with a custom command 00:02:20.569 [92/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:02:20.569 [93/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:02:20.829 [94/740] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:02:20.829 [95/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:02:20.829 [96/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:02:20.829 [97/740] Generating lib/rte_mbuf_def with a custom command 00:02:20.829 [98/740] Linking static target lib/librte_eal.a 00:02:20.829 [99/740] Generating lib/rte_mbuf_mingw with a custom command 00:02:20.829 [100/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:02:20.829 [101/740] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:02:20.829 [102/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:02:21.089 [103/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:02:21.089 [104/740] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:02:21.089 [105/740] Linking static target lib/librte_rcu.a 00:02:21.089 [106/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:02:21.089 [107/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:02:21.089 [108/740] Linking static target lib/librte_mempool.a 00:02:21.349 [109/740] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:02:21.349 [110/740] Linking static target lib/net/libnet_crc_avx512_lib.a 00:02:21.349 [111/740] Generating lib/rte_net_def with a custom command 00:02:21.349 [112/740] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:02:21.349 [113/740] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:02:21.349 [114/740] Generating lib/rte_net_mingw with a custom command 00:02:21.349 [115/740] Generating lib/rte_meter_def with a custom command 00:02:21.349 [116/740] Generating lib/rte_meter_mingw with a custom command 00:02:21.349 [117/740] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:02:21.349 [118/740] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:02:21.350 [119/740] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:02:21.350 [120/740] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:02:21.350 [121/740] Linking static target lib/librte_meter.a 00:02:21.615 [122/740] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:02:21.615 [123/740] Linking static target lib/librte_net.a 00:02:21.615 [124/740] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:02:21.615 [125/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:02:21.615 [126/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:02:21.615 [127/740] Linking static target lib/librte_mbuf.a 00:02:21.615 [128/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:02:21.876 [129/740] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:02:21.876 [130/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:02:21.876 [131/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:02:21.876 [132/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:02:21.876 [133/740] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:02:22.135 [134/740] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:22.135 [135/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:02:22.135 [136/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:02:22.135 [137/740] Generating lib/rte_ethdev_def with a custom command 00:02:22.135 [138/740] Generating lib/rte_ethdev_mingw with a custom command 00:02:22.135 [139/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:02:22.395 [140/740] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:02:22.395 [141/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:02:22.395 [142/740] Linking static target lib/librte_pci.a 00:02:22.395 [143/740] Generating lib/rte_pci_def with a custom command 00:02:22.395 [144/740] Generating lib/rte_pci_mingw with a custom command 00:02:22.395 [145/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:02:22.395 [146/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:02:22.395 [147/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:02:22.395 [148/740] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:22.395 [149/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:02:22.395 [150/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:02:22.395 [151/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:02:22.655 [152/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:02:22.655 [153/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:02:22.655 [154/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:02:22.655 [155/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:02:22.655 [156/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:02:22.655 [157/740] Generating lib/rte_cmdline_def with a custom command 00:02:22.655 [158/740] Generating lib/rte_cmdline_mingw with a custom command 00:02:22.655 [159/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:02:22.655 [160/740] Generating lib/rte_metrics_def with a custom command 00:02:22.655 [161/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:02:22.655 [162/740] Generating lib/rte_metrics_mingw with a custom command 00:02:22.655 [163/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:02:22.655 [164/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:02:22.655 [165/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:02:22.655 [166/740] Generating lib/rte_hash_def with a custom command 00:02:22.655 [167/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics.c.o 00:02:22.655 [168/740] Linking static target lib/librte_cmdline.a 00:02:22.655 [169/740] Generating lib/rte_hash_mingw with a custom command 00:02:22.914 [170/740] Generating lib/rte_timer_def with a custom command 00:02:22.915 [171/740] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:02:22.915 [172/740] Generating lib/rte_timer_mingw with a custom command 00:02:22.915 [173/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:02:22.915 [174/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics_telemetry.c.o 00:02:22.915 [175/740] Linking static target lib/librte_metrics.a 00:02:23.172 [176/740] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:02:23.172 [177/740] Linking static target lib/librte_timer.a 00:02:23.172 [178/740] Generating lib/metrics.sym_chk with a custom command (wrapped by meson to capture output) 00:02:23.429 [179/740] Compiling C object lib/librte_acl.a.p/acl_acl_gen.c.o 00:02:23.429 [180/740] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:02:23.429 [181/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_scalar.c.o 00:02:23.429 [182/740] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:02:23.429 [183/740] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:23.429 [184/740] Generating lib/rte_acl_def with a custom command 00:02:23.429 [185/740] Generating lib/rte_acl_mingw with a custom command 00:02:23.688 [186/740] Compiling C object lib/librte_acl.a.p/acl_rte_acl.c.o 00:02:23.688 [187/740] Generating lib/rte_bbdev_def with a custom command 00:02:23.688 [188/740] Compiling C object lib/librte_acl.a.p/acl_tb_mem.c.o 00:02:23.688 [189/740] Generating lib/rte_bbdev_mingw with a custom command 00:02:23.688 [190/740] Generating lib/rte_bitratestats_def with a custom command 00:02:23.688 [191/740] Generating lib/rte_bitratestats_mingw with a custom command 00:02:23.947 [192/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:02:23.947 [193/740] Linking static target lib/librte_ethdev.a 00:02:23.947 [194/740] Compiling C object lib/librte_bitratestats.a.p/bitratestats_rte_bitrate.c.o 00:02:23.947 [195/740] Linking static target lib/librte_bitratestats.a 00:02:24.206 [196/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf.c.o 00:02:24.206 [197/740] Compiling C object lib/librte_acl.a.p/acl_acl_bld.c.o 00:02:24.206 [198/740] Compiling C object lib/librte_bbdev.a.p/bbdev_rte_bbdev.c.o 00:02:24.206 [199/740] Linking static target lib/librte_bbdev.a 00:02:24.206 [200/740] Generating lib/bitratestats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:24.465 [201/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_dump.c.o 00:02:24.724 [202/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load.c.o 00:02:24.724 [203/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_exec.c.o 00:02:24.724 [204/740] Generating lib/bbdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:24.984 [205/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_stub.c.o 00:02:24.984 [206/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_sse.c.o 00:02:24.984 [207/740] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:02:24.984 [208/740] Linking static target lib/librte_hash.a 00:02:25.243 [209/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_pkt.c.o 00:02:25.243 [210/740] Generating lib/rte_bpf_def with a custom command 00:02:25.243 [211/740] Generating lib/rte_bpf_mingw with a custom command 00:02:25.243 [212/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load_elf.c.o 00:02:25.243 [213/740] Generating lib/rte_cfgfile_def with a custom command 00:02:25.520 [214/740] Generating lib/rte_cfgfile_mingw with a custom command 00:02:25.520 [215/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_convert.c.o 00:02:25.520 [216/740] Compiling C object lib/librte_cfgfile.a.p/cfgfile_rte_cfgfile.c.o 00:02:25.520 [217/740] Linking static target lib/librte_cfgfile.a 00:02:25.520 [218/740] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:02:25.520 [219/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_validate.c.o 00:02:25.520 [220/740] Generating lib/rte_compressdev_def with a custom command 00:02:25.520 [221/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx2.c.o 00:02:25.520 [222/740] Generating lib/rte_compressdev_mingw with a custom command 00:02:25.520 [223/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_jit_x86.c.o 00:02:25.520 [224/740] Linking static target lib/librte_bpf.a 00:02:25.780 [225/740] Generating lib/cfgfile.sym_chk with a custom command (wrapped by meson to capture output) 00:02:25.780 [226/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:02:25.780 [227/740] Generating lib/rte_cryptodev_def with a custom command 00:02:25.780 [228/740] Generating lib/rte_cryptodev_mingw with a custom command 00:02:25.780 [229/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:02:25.780 [230/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx512.c.o 00:02:25.780 [231/740] Generating lib/bpf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:25.780 [232/740] Linking static target lib/librte_acl.a 00:02:26.041 [233/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:02:26.041 [234/740] Linking static target lib/librte_compressdev.a 00:02:26.041 [235/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:02:26.041 [236/740] Generating lib/rte_distributor_def with a custom command 00:02:26.041 [237/740] Generating lib/rte_distributor_mingw with a custom command 00:02:26.041 [238/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:02:26.041 [239/740] Generating lib/rte_efd_def with a custom command 00:02:26.041 [240/740] Generating lib/rte_efd_mingw with a custom command 00:02:26.041 [241/740] Generating lib/acl.sym_chk with a custom command (wrapped by meson to capture output) 00:02:26.301 [242/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_match_sse.c.o 00:02:26.301 [243/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_single.c.o 00:02:26.301 [244/740] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:02:26.301 [245/740] Linking target lib/librte_eal.so.23.0 00:02:26.562 [246/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_private.c.o 00:02:26.562 [247/740] Generating symbol file lib/librte_eal.so.23.0.p/librte_eal.so.23.0.symbols 00:02:26.562 [248/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor.c.o 00:02:26.562 [249/740] Linking target lib/librte_ring.so.23.0 00:02:26.562 [250/740] Linking target lib/librte_meter.so.23.0 00:02:26.562 [251/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_trace_points.c.o 00:02:26.562 [252/740] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:26.562 [253/740] Generating symbol file lib/librte_ring.so.23.0.p/librte_ring.so.23.0.symbols 00:02:26.562 [254/740] Linking target lib/librte_pci.so.23.0 00:02:26.562 [255/740] Linking target lib/librte_timer.so.23.0 00:02:26.562 [256/740] Generating symbol file lib/librte_meter.so.23.0.p/librte_meter.so.23.0.symbols 00:02:26.822 [257/740] Linking target lib/librte_rcu.so.23.0 00:02:26.822 [258/740] Linking target lib/librte_mempool.so.23.0 00:02:26.822 [259/740] Generating symbol file lib/librte_timer.so.23.0.p/librte_timer.so.23.0.symbols 00:02:26.822 [260/740] Generating symbol file lib/librte_pci.so.23.0.p/librte_pci.so.23.0.symbols 00:02:26.822 [261/740] Generating symbol file lib/librte_rcu.so.23.0.p/librte_rcu.so.23.0.symbols 00:02:26.822 [262/740] Linking static target lib/librte_distributor.a 00:02:26.822 [263/740] Linking target lib/librte_cfgfile.so.23.0 00:02:26.822 [264/740] Linking target lib/librte_acl.so.23.0 00:02:26.822 [265/740] Generating symbol file lib/librte_mempool.so.23.0.p/librte_mempool.so.23.0.symbols 00:02:26.822 [266/740] Linking target lib/librte_mbuf.so.23.0 00:02:26.822 [267/740] Generating symbol file lib/librte_acl.so.23.0.p/librte_acl.so.23.0.symbols 00:02:27.083 [268/740] Generating symbol file lib/librte_mbuf.so.23.0.p/librte_mbuf.so.23.0.symbols 00:02:27.083 [269/740] Generating lib/distributor.sym_chk with a custom command (wrapped by meson to capture output) 00:02:27.083 [270/740] Linking target lib/librte_net.so.23.0 00:02:27.083 [271/740] Linking target lib/librte_bbdev.so.23.0 00:02:27.083 [272/740] Compiling C object lib/librte_efd.a.p/efd_rte_efd.c.o 00:02:27.083 [273/740] Linking target lib/librte_compressdev.so.23.0 00:02:27.083 [274/740] Linking static target lib/librte_efd.a 00:02:27.083 [275/740] Linking target lib/librte_distributor.so.23.0 00:02:27.083 [276/740] Generating symbol file lib/librte_net.so.23.0.p/librte_net.so.23.0.symbols 00:02:27.083 [277/740] Linking target lib/librte_cmdline.so.23.0 00:02:27.083 [278/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_ring.c.o 00:02:27.083 [279/740] Linking target lib/librte_hash.so.23.0 00:02:27.083 [280/740] Generating lib/rte_eventdev_def with a custom command 00:02:27.343 [281/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_crypto_adapter.c.o 00:02:27.343 [282/740] Generating lib/rte_eventdev_mingw with a custom command 00:02:27.343 [283/740] Generating lib/rte_gpudev_def with a custom command 00:02:27.343 [284/740] Generating lib/efd.sym_chk with a custom command (wrapped by meson to capture output) 00:02:27.343 [285/740] Generating lib/rte_gpudev_mingw with a custom command 00:02:27.343 [286/740] Generating symbol file lib/librte_hash.so.23.0.p/librte_hash.so.23.0.symbols 00:02:27.343 [287/740] Linking target lib/librte_efd.so.23.0 00:02:27.343 [288/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:02:27.343 [289/740] Linking static target lib/librte_cryptodev.a 00:02:27.603 [290/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_tx_adapter.c.o 00:02:27.863 [291/740] Compiling C object lib/librte_gro.a.p/gro_rte_gro.c.o 00:02:27.863 [292/740] Compiling C object lib/librte_gro.a.p/gro_gro_tcp4.c.o 00:02:27.863 [293/740] Generating lib/rte_gro_def with a custom command 00:02:27.863 [294/740] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:27.863 [295/740] Generating lib/rte_gro_mingw with a custom command 00:02:27.863 [296/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_timer_adapter.c.o 00:02:27.863 [297/740] Compiling C object lib/librte_gro.a.p/gro_gro_udp4.c.o 00:02:27.863 [298/740] Linking target lib/librte_ethdev.so.23.0 00:02:27.863 [299/740] Compiling C object lib/librte_gpudev.a.p/gpudev_gpudev.c.o 00:02:27.863 [300/740] Linking static target lib/librte_gpudev.a 00:02:27.863 [301/740] Generating symbol file lib/librte_ethdev.so.23.0.p/librte_ethdev.so.23.0.symbols 00:02:28.123 [302/740] Linking target lib/librte_metrics.so.23.0 00:02:28.123 [303/740] Generating symbol file lib/librte_metrics.so.23.0.p/librte_metrics.so.23.0.symbols 00:02:28.123 [304/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_eventdev.c.o 00:02:28.123 [305/740] Linking target lib/librte_bitratestats.so.23.0 00:02:28.123 [306/740] Compiling C object lib/librte_gso.a.p/gso_gso_udp4.c.o 00:02:28.123 [307/740] Compiling C object lib/librte_gso.a.p/gso_gso_tcp4.c.o 00:02:28.123 [308/740] Linking target lib/librte_bpf.so.23.0 00:02:28.382 [309/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_tcp4.c.o 00:02:28.382 [310/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_udp4.c.o 00:02:28.382 [311/740] Linking static target lib/librte_gro.a 00:02:28.382 [312/740] Generating lib/rte_gso_def with a custom command 00:02:28.382 [313/740] Generating lib/rte_gso_mingw with a custom command 00:02:28.382 [314/740] Generating symbol file lib/librte_bpf.so.23.0.p/librte_bpf.so.23.0.symbols 00:02:28.382 [315/740] Compiling C object lib/librte_gso.a.p/gso_gso_common.c.o 00:02:28.382 [316/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_rx_adapter.c.o 00:02:28.382 [317/740] Linking static target lib/librte_eventdev.a 00:02:28.382 [318/740] Generating lib/gro.sym_chk with a custom command (wrapped by meson to capture output) 00:02:28.641 [319/740] Linking target lib/librte_gro.so.23.0 00:02:28.641 [320/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_udp4.c.o 00:02:28.641 [321/740] Compiling C object lib/librte_gso.a.p/gso_rte_gso.c.o 00:02:28.641 [322/740] Generating lib/gpudev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:28.641 [323/740] Linking target lib/librte_gpudev.so.23.0 00:02:28.641 [324/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_tcp4.c.o 00:02:28.641 [325/740] Linking static target lib/librte_gso.a 00:02:28.641 [326/740] Generating lib/rte_ip_frag_def with a custom command 00:02:28.641 [327/740] Generating lib/rte_ip_frag_mingw with a custom command 00:02:28.901 [328/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_reassembly.c.o 00:02:28.901 [329/740] Compiling C object lib/librte_jobstats.a.p/jobstats_rte_jobstats.c.o 00:02:28.901 [330/740] Generating lib/gso.sym_chk with a custom command (wrapped by meson to capture output) 00:02:28.901 [331/740] Linking static target lib/librte_jobstats.a 00:02:28.901 [332/740] Generating lib/rte_jobstats_def with a custom command 00:02:28.901 [333/740] Generating lib/rte_jobstats_mingw with a custom command 00:02:28.901 [334/740] Linking target lib/librte_gso.so.23.0 00:02:28.901 [335/740] Generating lib/rte_latencystats_def with a custom command 00:02:28.901 [336/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_fragmentation.c.o 00:02:28.901 [337/740] Generating lib/rte_latencystats_mingw with a custom command 00:02:28.901 [338/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_reassembly.c.o 00:02:28.901 [339/740] Generating lib/rte_lpm_def with a custom command 00:02:28.901 [340/740] Generating lib/rte_lpm_mingw with a custom command 00:02:29.160 [341/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_fragmentation.c.o 00:02:29.160 [342/740] Generating lib/jobstats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.160 [343/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ip_frag_common.c.o 00:02:29.160 [344/740] Linking target lib/librte_jobstats.so.23.0 00:02:29.160 [345/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_ip_frag_internal.c.o 00:02:29.160 [346/740] Linking static target lib/librte_ip_frag.a 00:02:29.420 [347/740] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.420 [348/740] Compiling C object lib/librte_latencystats.a.p/latencystats_rte_latencystats.c.o 00:02:29.420 [349/740] Linking static target lib/librte_latencystats.a 00:02:29.420 [350/740] Linking target lib/librte_cryptodev.so.23.0 00:02:29.420 [351/740] Generating lib/ip_frag.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.420 [352/740] Linking target lib/librte_ip_frag.so.23.0 00:02:29.420 [353/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm.c.o 00:02:29.420 [354/740] Generating symbol file lib/librte_cryptodev.so.23.0.p/librte_cryptodev.so.23.0.symbols 00:02:29.420 [355/740] Compiling C object lib/librte_member.a.p/member_rte_member.c.o 00:02:29.420 [356/740] Compiling C object lib/member/libsketch_avx512_tmp.a.p/rte_member_sketch_avx512.c.o 00:02:29.420 [357/740] Generating lib/rte_member_def with a custom command 00:02:29.420 [358/740] Generating lib/rte_member_mingw with a custom command 00:02:29.420 [359/740] Linking static target lib/member/libsketch_avx512_tmp.a 00:02:29.420 [360/740] Generating lib/latencystats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.420 [361/740] Generating symbol file lib/librte_ip_frag.so.23.0.p/librte_ip_frag.so.23.0.symbols 00:02:29.420 [362/740] Generating lib/rte_pcapng_def with a custom command 00:02:29.420 [363/740] Generating lib/rte_pcapng_mingw with a custom command 00:02:29.420 [364/740] Linking target lib/librte_latencystats.so.23.0 00:02:29.680 [365/740] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:02:29.680 [366/740] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:02:29.680 [367/740] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:02:29.680 [368/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm6.c.o 00:02:29.680 [369/740] Linking static target lib/librte_lpm.a 00:02:29.680 [370/740] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:02:29.940 [371/740] Compiling C object lib/librte_member.a.p/member_rte_member_ht.c.o 00:02:29.940 [372/740] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:02:29.940 [373/740] Compiling C object lib/librte_power.a.p/power_rte_power_empty_poll.c.o 00:02:29.940 [374/740] Compiling C object lib/librte_member.a.p/member_rte_member_vbf.c.o 00:02:29.940 [375/740] Generating lib/rte_power_def with a custom command 00:02:29.940 [376/740] Generating lib/rte_power_mingw with a custom command 00:02:29.940 [377/740] Generating lib/eventdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.940 [378/740] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:02:29.940 [379/740] Generating lib/rte_rawdev_def with a custom command 00:02:29.940 [380/740] Generating lib/rte_rawdev_mingw with a custom command 00:02:29.940 [381/740] Generating lib/lpm.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.940 [382/740] Linking target lib/librte_eventdev.so.23.0 00:02:29.940 [383/740] Generating lib/rte_regexdev_def with a custom command 00:02:29.940 [384/740] Linking target lib/librte_lpm.so.23.0 00:02:30.200 [385/740] Generating lib/rte_regexdev_mingw with a custom command 00:02:30.200 [386/740] Generating symbol file lib/librte_eventdev.so.23.0.p/librte_eventdev.so.23.0.symbols 00:02:30.200 [387/740] Generating lib/rte_dmadev_def with a custom command 00:02:30.200 [388/740] Generating symbol file lib/librte_lpm.so.23.0.p/librte_lpm.so.23.0.symbols 00:02:30.200 [389/740] Generating lib/rte_dmadev_mingw with a custom command 00:02:30.200 [390/740] Compiling C object lib/librte_pcapng.a.p/pcapng_rte_pcapng.c.o 00:02:30.200 [391/740] Linking static target lib/librte_pcapng.a 00:02:30.200 [392/740] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:02:30.200 [393/740] Generating lib/rte_rib_def with a custom command 00:02:30.200 [394/740] Generating lib/rte_rib_mingw with a custom command 00:02:30.200 [395/740] Compiling C object lib/librte_rawdev.a.p/rawdev_rte_rawdev.c.o 00:02:30.200 [396/740] Compiling C object lib/librte_power.a.p/power_rte_power_intel_uncore.c.o 00:02:30.200 [397/740] Linking static target lib/librte_rawdev.a 00:02:30.200 [398/740] Generating lib/rte_reorder_def with a custom command 00:02:30.460 [399/740] Generating lib/rte_reorder_mingw with a custom command 00:02:30.460 [400/740] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:02:30.460 [401/740] Linking static target lib/librte_dmadev.a 00:02:30.460 [402/740] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:02:30.460 [403/740] Linking static target lib/librte_power.a 00:02:30.460 [404/740] Generating lib/pcapng.sym_chk with a custom command (wrapped by meson to capture output) 00:02:30.460 [405/740] Compiling C object lib/librte_regexdev.a.p/regexdev_rte_regexdev.c.o 00:02:30.460 [406/740] Linking static target lib/librte_regexdev.a 00:02:30.460 [407/740] Linking target lib/librte_pcapng.so.23.0 00:02:30.460 [408/740] Generating symbol file lib/librte_pcapng.so.23.0.p/librte_pcapng.so.23.0.symbols 00:02:30.721 [409/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib.c.o 00:02:30.721 [410/740] Generating lib/rawdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:30.721 [411/740] Compiling C object lib/librte_sched.a.p/sched_rte_red.c.o 00:02:30.721 [412/740] Linking target lib/librte_rawdev.so.23.0 00:02:30.721 [413/740] Generating lib/rte_sched_def with a custom command 00:02:30.721 [414/740] Generating lib/rte_sched_mingw with a custom command 00:02:30.721 [415/740] Compiling C object lib/librte_member.a.p/member_rte_member_sketch.c.o 00:02:30.721 [416/740] Compiling C object lib/librte_sched.a.p/sched_rte_approx.c.o 00:02:30.721 [417/740] Linking static target lib/librte_member.a 00:02:30.721 [418/740] Compiling C object lib/librte_sched.a.p/sched_rte_pie.c.o 00:02:30.721 [419/740] Generating lib/rte_security_def with a custom command 00:02:30.721 [420/740] Generating lib/rte_security_mingw with a custom command 00:02:30.721 [421/740] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:02:30.721 [422/740] Linking static target lib/librte_reorder.a 00:02:30.981 [423/740] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:30.981 [424/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib6.c.o 00:02:30.981 [425/740] Linking static target lib/librte_rib.a 00:02:30.981 [426/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_std.c.o 00:02:30.981 [427/740] Linking target lib/librte_dmadev.so.23.0 00:02:30.981 [428/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack.c.o 00:02:30.981 [429/740] Generating lib/rte_stack_def with a custom command 00:02:30.981 [430/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_lf.c.o 00:02:30.981 [431/740] Generating lib/rte_stack_mingw with a custom command 00:02:30.981 [432/740] Linking static target lib/librte_stack.a 00:02:30.981 [433/740] Generating symbol file lib/librte_dmadev.so.23.0.p/librte_dmadev.so.23.0.symbols 00:02:30.981 [434/740] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:02:30.981 [435/740] Generating lib/member.sym_chk with a custom command (wrapped by meson to capture output) 00:02:30.981 [436/740] Linking target lib/librte_reorder.so.23.0 00:02:30.981 [437/740] Generating lib/regexdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:30.981 [438/740] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:02:30.981 [439/740] Linking target lib/librte_member.so.23.0 00:02:31.241 [440/740] Generating lib/stack.sym_chk with a custom command (wrapped by meson to capture output) 00:02:31.241 [441/740] Linking target lib/librte_regexdev.so.23.0 00:02:31.241 [442/740] Linking target lib/librte_stack.so.23.0 00:02:31.241 [443/740] Generating lib/rib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:31.241 [444/740] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:02:31.241 [445/740] Linking target lib/librte_rib.so.23.0 00:02:31.241 [446/740] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:02:31.241 [447/740] Linking target lib/librte_power.so.23.0 00:02:31.241 [448/740] Linking static target lib/librte_security.a 00:02:31.501 [449/740] Generating symbol file lib/librte_rib.so.23.0.p/librte_rib.so.23.0.symbols 00:02:31.501 [450/740] Generating lib/rte_vhost_def with a custom command 00:02:31.501 [451/740] Generating lib/rte_vhost_mingw with a custom command 00:02:31.501 [452/740] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:02:31.501 [453/740] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:02:31.501 [454/740] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:02:31.501 [455/740] Linking target lib/librte_security.so.23.0 00:02:31.761 [456/740] Compiling C object lib/librte_sched.a.p/sched_rte_sched.c.o 00:02:31.761 [457/740] Linking static target lib/librte_sched.a 00:02:31.761 [458/740] Generating symbol file lib/librte_security.so.23.0.p/librte_security.so.23.0.symbols 00:02:31.761 [459/740] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:02:32.020 [460/740] Generating lib/sched.sym_chk with a custom command (wrapped by meson to capture output) 00:02:32.020 [461/740] Compiling C object lib/librte_ipsec.a.p/ipsec_sa.c.o 00:02:32.020 [462/740] Linking target lib/librte_sched.so.23.0 00:02:32.020 [463/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ses.c.o 00:02:32.020 [464/740] Generating lib/rte_ipsec_def with a custom command 00:02:32.020 [465/740] Generating lib/rte_ipsec_mingw with a custom command 00:02:32.020 [466/740] Generating symbol file lib/librte_sched.so.23.0.p/librte_sched.so.23.0.symbols 00:02:32.279 [467/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib.c.o 00:02:32.279 [468/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:02:32.279 [469/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:02:32.279 [470/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_sad.c.o 00:02:32.279 [471/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_telemetry.c.o 00:02:32.279 [472/740] Generating lib/rte_fib_def with a custom command 00:02:32.538 [473/740] Generating lib/rte_fib_mingw with a custom command 00:02:32.538 [474/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib6.c.o 00:02:32.797 [475/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_outb.c.o 00:02:32.797 [476/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8_avx512.c.o 00:02:32.797 [477/740] Compiling C object lib/librte_fib.a.p/fib_trie_avx512.c.o 00:02:32.797 [478/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_inb.c.o 00:02:32.797 [479/740] Linking static target lib/librte_ipsec.a 00:02:33.056 [480/740] Compiling C object lib/librte_fib.a.p/fib_trie.c.o 00:02:33.056 [481/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8.c.o 00:02:33.056 [482/740] Linking static target lib/librte_fib.a 00:02:33.056 [483/740] Generating lib/ipsec.sym_chk with a custom command (wrapped by meson to capture output) 00:02:33.056 [484/740] Compiling C object lib/librte_port.a.p/port_rte_port_fd.c.o 00:02:33.056 [485/740] Compiling C object lib/librte_port.a.p/port_rte_port_ethdev.c.o 00:02:33.056 [486/740] Linking target lib/librte_ipsec.so.23.0 00:02:33.315 [487/740] Generating lib/fib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:33.315 [488/740] Compiling C object lib/librte_port.a.p/port_rte_port_frag.c.o 00:02:33.315 [489/740] Compiling C object lib/librte_port.a.p/port_rte_port_ras.c.o 00:02:33.315 [490/740] Linking target lib/librte_fib.so.23.0 00:02:33.315 [491/740] Compiling C object lib/librte_port.a.p/port_rte_port_sched.c.o 00:02:33.884 [492/740] Compiling C object lib/librte_port.a.p/port_rte_port_sym_crypto.c.o 00:02:33.884 [493/740] Generating lib/rte_port_def with a custom command 00:02:33.884 [494/740] Generating lib/rte_port_mingw with a custom command 00:02:33.884 [495/740] Compiling C object lib/librte_port.a.p/port_rte_port_source_sink.c.o 00:02:33.884 [496/740] Compiling C object lib/librte_port.a.p/port_rte_port_eventdev.c.o 00:02:33.884 [497/740] Generating lib/rte_pdump_def with a custom command 00:02:33.884 [498/740] Generating lib/rte_pdump_mingw with a custom command 00:02:33.884 [499/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ethdev.c.o 00:02:33.884 [500/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_fd.c.o 00:02:34.144 [501/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_source_sink.c.o 00:02:34.144 [502/740] Compiling C object lib/librte_table.a.p/table_rte_swx_keycmp.c.o 00:02:34.144 [503/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_learner.c.o 00:02:34.144 [504/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_em.c.o 00:02:34.144 [505/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ring.c.o 00:02:34.403 [506/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_selector.c.o 00:02:34.403 [507/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_wm.c.o 00:02:34.403 [508/740] Compiling C object lib/librte_port.a.p/port_rte_port_ring.c.o 00:02:34.403 [509/740] Linking static target lib/librte_port.a 00:02:34.403 [510/740] Compiling C object lib/librte_table.a.p/table_rte_table_array.c.o 00:02:34.663 [511/740] Compiling C object lib/librte_table.a.p/table_rte_table_acl.c.o 00:02:34.663 [512/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_cuckoo.c.o 00:02:34.663 [513/740] Compiling C object lib/librte_pdump.a.p/pdump_rte_pdump.c.o 00:02:34.663 [514/740] Linking static target lib/librte_pdump.a 00:02:34.923 [515/740] Generating lib/port.sym_chk with a custom command (wrapped by meson to capture output) 00:02:34.923 [516/740] Linking target lib/librte_port.so.23.0 00:02:34.923 [517/740] Generating lib/pdump.sym_chk with a custom command (wrapped by meson to capture output) 00:02:34.923 [518/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm.c.o 00:02:34.923 [519/740] Linking target lib/librte_pdump.so.23.0 00:02:34.923 [520/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_ext.c.o 00:02:34.923 [521/740] Generating symbol file lib/librte_port.so.23.0.p/librte_port.so.23.0.symbols 00:02:34.923 [522/740] Generating lib/rte_table_def with a custom command 00:02:35.182 [523/740] Generating lib/rte_table_mingw with a custom command 00:02:35.182 [524/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key8.c.o 00:02:35.182 [525/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key16.c.o 00:02:35.441 [526/740] Compiling C object lib/librte_table.a.p/table_rte_table_stub.c.o 00:02:35.441 [527/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm_ipv6.c.o 00:02:35.441 [528/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_lru.c.o 00:02:35.441 [529/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key32.c.o 00:02:35.441 [530/740] Generating lib/rte_pipeline_def with a custom command 00:02:35.441 [531/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_port_in_action.c.o 00:02:35.441 [532/740] Linking static target lib/librte_table.a 00:02:35.441 [533/740] Generating lib/rte_pipeline_mingw with a custom command 00:02:35.700 [534/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_pipeline.c.o 00:02:35.959 [535/740] Compiling C object lib/librte_graph.a.p/graph_node.c.o 00:02:35.959 [536/740] Compiling C object lib/librte_graph.a.p/graph_graph.c.o 00:02:35.959 [537/740] Generating lib/table.sym_chk with a custom command (wrapped by meson to capture output) 00:02:35.959 [538/740] Linking target lib/librte_table.so.23.0 00:02:36.219 [539/740] Compiling C object lib/librte_graph.a.p/graph_graph_ops.c.o 00:02:36.219 [540/740] Generating symbol file lib/librte_table.so.23.0.p/librte_table.so.23.0.symbols 00:02:36.219 [541/740] Generating lib/rte_graph_def with a custom command 00:02:36.219 [542/740] Generating lib/rte_graph_mingw with a custom command 00:02:36.219 [543/740] Compiling C object lib/librte_graph.a.p/graph_graph_debug.c.o 00:02:36.479 [544/740] Compiling C object lib/librte_graph.a.p/graph_graph_stats.c.o 00:02:36.479 [545/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ctl.c.o 00:02:36.479 [546/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:02:36.479 [547/740] Compiling C object lib/librte_graph.a.p/graph_graph_populate.c.o 00:02:36.479 [548/740] Linking static target lib/librte_graph.a 00:02:36.739 [549/740] Compiling C object lib/librte_node.a.p/node_ethdev_ctrl.c.o 00:02:36.739 [550/740] Compiling C object lib/librte_node.a.p/node_ethdev_rx.c.o 00:02:36.739 [551/740] Compiling C object lib/librte_node.a.p/node_null.c.o 00:02:36.739 [552/740] Compiling C object lib/librte_node.a.p/node_ethdev_tx.c.o 00:02:36.999 [553/740] Compiling C object lib/librte_node.a.p/node_log.c.o 00:02:36.999 [554/740] Generating lib/rte_node_def with a custom command 00:02:36.999 [555/740] Generating lib/rte_node_mingw with a custom command 00:02:37.259 [556/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline_spec.c.o 00:02:37.259 [557/740] Compiling C object lib/librte_node.a.p/node_pkt_drop.c.o 00:02:37.259 [558/740] Generating lib/graph.sym_chk with a custom command (wrapped by meson to capture output) 00:02:37.259 [559/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:02:37.259 [560/740] Linking target lib/librte_graph.so.23.0 00:02:37.259 [561/740] Compiling C object lib/librte_node.a.p/node_ip4_lookup.c.o 00:02:37.259 [562/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:02:37.259 [563/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:02:37.259 [564/740] Generating drivers/rte_bus_pci_def with a custom command 00:02:37.259 [565/740] Generating symbol file lib/librte_graph.so.23.0.p/librte_graph.so.23.0.symbols 00:02:37.259 [566/740] Generating drivers/rte_bus_pci_mingw with a custom command 00:02:37.518 [567/740] Compiling C object lib/librte_node.a.p/node_pkt_cls.c.o 00:02:37.518 [568/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:02:37.518 [569/740] Generating drivers/rte_bus_vdev_def with a custom command 00:02:37.518 [570/740] Generating drivers/rte_bus_vdev_mingw with a custom command 00:02:37.518 [571/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:02:37.518 [572/740] Generating drivers/rte_mempool_ring_def with a custom command 00:02:37.518 [573/740] Compiling C object lib/librte_node.a.p/node_ip4_rewrite.c.o 00:02:37.518 [574/740] Generating drivers/rte_mempool_ring_mingw with a custom command 00:02:37.518 [575/740] Linking static target lib/librte_node.a 00:02:37.518 [576/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:02:37.518 [577/740] Linking static target drivers/libtmp_rte_bus_vdev.a 00:02:37.518 [578/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:02:37.777 [579/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:02:37.777 [580/740] Generating lib/node.sym_chk with a custom command (wrapped by meson to capture output) 00:02:37.777 [581/740] Linking static target drivers/libtmp_rte_bus_pci.a 00:02:37.777 [582/740] Linking target lib/librte_node.so.23.0 00:02:37.777 [583/740] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:02:37.777 [584/740] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:37.777 [585/740] Linking static target drivers/librte_bus_vdev.a 00:02:37.778 [586/740] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:02:37.778 [587/740] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:37.778 [588/740] Linking static target drivers/librte_bus_pci.a 00:02:38.037 [589/740] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:38.037 [590/740] Compiling C object drivers/librte_bus_pci.so.23.0.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:38.037 [591/740] Compiling C object drivers/librte_bus_vdev.so.23.0.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:38.037 [592/740] Linking target drivers/librte_bus_vdev.so.23.0 00:02:38.037 [593/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_diag.c.o 00:02:38.037 [594/740] Generating symbol file drivers/librte_bus_vdev.so.23.0.p/librte_bus_vdev.so.23.0.symbols 00:02:38.037 [595/740] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:38.298 [596/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_dcb.c.o 00:02:38.298 [597/740] Linking target drivers/librte_bus_pci.so.23.0 00:02:38.298 [598/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_adminq.c.o 00:02:38.298 [599/740] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:02:38.298 [600/740] Linking static target drivers/libtmp_rte_mempool_ring.a 00:02:38.298 [601/740] Generating symbol file drivers/librte_bus_pci.so.23.0.p/librte_bus_pci.so.23.0.symbols 00:02:38.558 [602/740] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:02:38.558 [603/740] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:38.558 [604/740] Linking static target drivers/librte_mempool_ring.a 00:02:38.558 [605/740] Compiling C object drivers/librte_mempool_ring.so.23.0.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:38.559 [606/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_hmc.c.o 00:02:38.559 [607/740] Linking target drivers/librte_mempool_ring.so.23.0 00:02:38.828 [608/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_lan_hmc.c.o 00:02:39.118 [609/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_nvm.c.o 00:02:39.118 [610/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_common.c.o 00:02:39.397 [611/740] Linking static target drivers/net/i40e/base/libi40e_base.a 00:02:39.397 [612/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_pf.c.o 00:02:39.966 [613/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_fdir.c.o 00:02:39.966 [614/740] Compiling C object drivers/net/i40e/libi40e_avx512_lib.a.p/i40e_rxtx_vec_avx512.c.o 00:02:39.966 [615/740] Linking static target drivers/net/i40e/libi40e_avx512_lib.a 00:02:39.966 [616/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_tm.c.o 00:02:39.966 [617/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_hash.c.o 00:02:39.966 [618/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_flow.c.o 00:02:40.226 [619/740] Generating drivers/rte_net_i40e_def with a custom command 00:02:40.226 [620/740] Generating drivers/rte_net_i40e_mingw with a custom command 00:02:40.226 [621/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_vf_representor.c.o 00:02:40.795 [622/740] Compiling C object app/dpdk-dumpcap.p/dumpcap_main.c.o 00:02:40.795 [623/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline.c.o 00:02:41.055 [624/740] Compiling C object app/dpdk-pdump.p/pdump_main.c.o 00:02:41.314 [625/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_main.c.o 00:02:41.314 [626/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx.c.o 00:02:41.314 [627/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_sse.c.o 00:02:41.314 [628/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_rte_pmd_i40e.c.o 00:02:41.314 [629/740] Compiling C object app/dpdk-test-acl.p/test-acl_main.c.o 00:02:41.314 [630/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_commands.c.o 00:02:41.314 [631/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_avx2.c.o 00:02:41.314 [632/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_cmdline_test.c.o 00:02:41.573 [633/740] Compiling C object app/dpdk-proc-info.p/proc-info_main.c.o 00:02:41.833 [634/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_options_parse.c.o 00:02:41.833 [635/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_ethdev.c.o 00:02:41.833 [636/740] Linking static target drivers/libtmp_rte_net_i40e.a 00:02:42.092 [637/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_throughput.c.o 00:02:42.092 [638/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev.c.o 00:02:42.092 [639/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_vector.c.o 00:02:42.351 [640/740] Generating drivers/rte_net_i40e.pmd.c with a custom command 00:02:42.351 [641/740] Compiling C object drivers/librte_net_i40e.a.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:02:42.351 [642/740] Linking static target drivers/librte_net_i40e.a 00:02:42.351 [643/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_common.c.o 00:02:42.351 [644/740] Compiling C object drivers/librte_net_i40e.so.23.0.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:02:42.351 [645/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_cyclecount.c.o 00:02:42.610 [646/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_main.c.o 00:02:42.610 [647/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_verify.c.o 00:02:42.870 [648/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_common.c.o 00:02:42.870 [649/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_options_parsing.c.o 00:02:42.870 [650/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_ops.c.o 00:02:42.870 [651/740] Generating drivers/rte_net_i40e.sym_chk with a custom command (wrapped by meson to capture output) 00:02:42.870 [652/740] Linking target drivers/librte_net_i40e.so.23.0 00:02:43.130 [653/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vectors.c.o 00:02:43.390 [654/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_latency.c.o 00:02:43.390 [655/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vector_parsing.c.o 00:02:43.390 [656/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_pmd_cyclecount.c.o 00:02:43.390 [657/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_throughput.c.o 00:02:43.390 [658/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_test.c.o 00:02:43.390 [659/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_verify.c.o 00:02:43.390 [660/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_main.c.o 00:02:43.655 [661/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_parser.c.o 00:02:43.655 [662/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_main.c.o 00:02:43.915 [663/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_options.c.o 00:02:43.915 [664/740] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:02:43.915 [665/740] Linking static target lib/librte_vhost.a 00:02:43.915 [666/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_common.c.o 00:02:43.915 [667/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_atq.c.o 00:02:44.174 [668/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_queue.c.o 00:02:44.433 [669/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_atq.c.o 00:02:44.433 [670/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_queue.c.o 00:02:44.693 [671/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_atq.c.o 00:02:44.693 [672/740] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:02:44.693 [673/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_common.c.o 00:02:44.693 [674/740] Linking target lib/librte_vhost.so.23.0 00:02:44.953 [675/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_flow_gen.c.o 00:02:44.953 [676/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_queue.c.o 00:02:44.953 [677/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_actions_gen.c.o 00:02:44.953 [678/740] Compiling C object app/dpdk-test-fib.p/test-fib_main.c.o 00:02:45.212 [679/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_items_gen.c.o 00:02:45.212 [680/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_common.c.o 00:02:45.212 [681/740] Compiling C object app/dpdk-test-gpudev.p/test-gpudev_main.c.o 00:02:45.212 [682/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_config.c.o 00:02:45.212 [683/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_init.c.o 00:02:45.212 [684/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_main.c.o 00:02:45.471 [685/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_acl.c.o 00:02:45.471 [686/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm.c.o 00:02:45.471 [687/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_hash.c.o 00:02:45.757 [688/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm_ipv6.c.o 00:02:45.757 [689/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_stub.c.o 00:02:45.757 [690/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_main.c.o 00:02:46.016 [691/740] Compiling C object app/dpdk-testpmd.p/test-pmd_5tswap.c.o 00:02:46.016 [692/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_runtime.c.o 00:02:46.016 [693/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmd_flex_item.c.o 00:02:46.016 [694/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_perf.c.o 00:02:46.274 [695/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_mtr.c.o 00:02:46.274 [696/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_tm.c.o 00:02:46.532 [697/740] Compiling C object app/dpdk-testpmd.p/test-pmd_flowgen.c.o 00:02:46.532 [698/740] Compiling C object app/dpdk-testpmd.p/test-pmd_icmpecho.c.o 00:02:46.532 [699/740] Compiling C object app/dpdk-testpmd.p/test-pmd_ieee1588fwd.c.o 00:02:46.790 [700/740] Compiling C object app/dpdk-testpmd.p/test-pmd_iofwd.c.o 00:02:46.790 [701/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macfwd.c.o 00:02:47.048 [702/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macswap.c.o 00:02:47.048 [703/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline.c.o 00:02:47.306 [704/740] Compiling C object app/dpdk-testpmd.p/test-pmd_csumonly.c.o 00:02:47.306 [705/740] Compiling C object app/dpdk-testpmd.p/test-pmd_rxonly.c.o 00:02:47.306 [706/740] Compiling C object app/dpdk-testpmd.p/test-pmd_parameters.c.o 00:02:47.306 [707/740] Compiling C object app/dpdk-testpmd.p/test-pmd_shared_rxq_fwd.c.o 00:02:47.601 [708/740] Compiling C object app/dpdk-testpmd.p/test-pmd_bpf_cmd.c.o 00:02:47.860 [709/740] Compiling C object app/dpdk-testpmd.p/test-pmd_util.c.o 00:02:47.860 [710/740] Compiling C object app/dpdk-testpmd.p/.._drivers_net_i40e_i40e_testpmd.c.o 00:02:48.120 [711/740] Compiling C object app/dpdk-testpmd.p/test-pmd_config.c.o 00:02:48.120 [712/740] Compiling C object app/dpdk-test-sad.p/test-sad_main.c.o 00:02:48.120 [713/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_flow.c.o 00:02:48.120 [714/740] Compiling C object app/dpdk-testpmd.p/test-pmd_txonly.c.o 00:02:48.120 [715/740] Compiling C object app/dpdk-test-regex.p/test-regex_main.c.o 00:02:48.379 [716/740] Compiling C object app/dpdk-test-security-perf.p/test-security-perf_test_security_perf.c.o 00:02:48.639 [717/740] Compiling C object app/dpdk-testpmd.p/test-pmd_testpmd.c.o 00:02:48.639 [718/740] Compiling C object app/dpdk-testpmd.p/test-pmd_noisy_vnf.c.o 00:02:48.899 [719/740] Compiling C object app/dpdk-test-security-perf.p/test_test_cryptodev_security_ipsec.c.o 00:02:49.159 [720/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_table_action.c.o 00:02:49.159 [721/740] Linking static target lib/librte_pipeline.a 00:02:49.419 [722/740] Linking target app/dpdk-test-cmdline 00:02:49.419 [723/740] Linking target app/dpdk-pdump 00:02:49.419 [724/740] Linking target app/dpdk-test-bbdev 00:02:49.419 [725/740] Linking target app/dpdk-test-compress-perf 00:02:49.419 [726/740] Linking target app/dpdk-test-crypto-perf 00:02:49.678 [727/740] Linking target app/dpdk-test-acl 00:02:49.678 [728/740] Linking target app/dpdk-dumpcap 00:02:49.678 [729/740] Linking target app/dpdk-proc-info 00:02:49.678 [730/740] Linking target app/dpdk-test-eventdev 00:02:49.938 [731/740] Linking target app/dpdk-test-gpudev 00:02:49.938 [732/740] Linking target app/dpdk-test-fib 00:02:49.938 [733/740] Linking target app/dpdk-test-flow-perf 00:02:49.938 [734/740] Linking target app/dpdk-test-pipeline 00:02:49.938 [735/740] Linking target app/dpdk-test-sad 00:02:49.938 [736/740] Linking target app/dpdk-test-regex 00:02:49.938 [737/740] Linking target app/dpdk-testpmd 00:02:49.938 [738/740] Linking target app/dpdk-test-security-perf 00:02:54.144 [739/740] Generating lib/pipeline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.144 [740/740] Linking target lib/librte_pipeline.so.23.0 00:02:54.144 05:30:27 build_native_dpdk -- common/autobuild_common.sh@201 -- $ uname -s 00:02:54.144 05:30:27 build_native_dpdk -- common/autobuild_common.sh@201 -- $ [[ Linux == \F\r\e\e\B\S\D ]] 00:02:54.144 05:30:27 build_native_dpdk -- common/autobuild_common.sh@214 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 install 00:02:54.144 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:02:54.144 [0/1] Installing files. 00:02:54.406 Installing subdir /home/vagrant/spdk_repo/dpdk/examples to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/README to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/dummy.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t1.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t2.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t3.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/common/pkt_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/common/altivec/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/altivec 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/common/neon/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/neon 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/common/sse/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/sse 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/dmafwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:54.406 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_tx.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_aes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ccm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_cmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ecdsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_gcm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_hmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_rsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_sha.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_tdes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_xts.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/flow_classify.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/ipv4_rules_file.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/flow_blocks.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.407 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/firewall.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow_crypto.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/kni.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route_ecmp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/tap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep0.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep1.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipip.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_process.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/rt.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp4.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp6.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/bypass_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs_secgw.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/data_rxtx.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/linux_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/load_env.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/run_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.408 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_ipv6opts.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_null_header_reconstruct.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/l2fwd-cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl_scalar.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.409 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_sequential.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_fib.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_route.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/client.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/shared 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/ntb_fwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:54.410 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ethdev.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_group_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_routing_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/packet.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/pcap.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/ptpclient.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/app_thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cmdline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.411 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_ov.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_pie.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_red.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/stats.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/node.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/shared 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/basicfwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/vdpa_blk_compact.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/virtio_net.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk_spec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk_compat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_nop.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_x86.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:02:54.412 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:02:54.412 Installing lib/librte_kvargs.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_telemetry.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_eal.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_rcu.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_mempool.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_mbuf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.413 Installing lib/librte_net.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_meter.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_ethdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_cmdline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_metrics.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_hash.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_timer.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_acl.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_bbdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_bitratestats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_bpf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_cfgfile.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_compressdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_cryptodev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_distributor.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_efd.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_eventdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_gpudev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_gro.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_gso.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_ip_frag.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_jobstats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_latencystats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_lpm.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_member.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_pcapng.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_power.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_rawdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_regexdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_dmadev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_rib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_reorder.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_sched.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_security.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_stack.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_vhost.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_ipsec.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_fib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_port.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_pdump.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_table.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_pipeline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_graph.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_node.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing lib/librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing drivers/librte_bus_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing drivers/librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:54.677 Installing drivers/librte_bus_vdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing drivers/librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:54.677 Installing drivers/librte_mempool_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing drivers/librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:54.677 Installing drivers/librte_net_i40e.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:54.677 Installing drivers/librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:54.677 Installing app/dpdk-dumpcap to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-pdump to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-proc-info to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-test-acl to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-test-bbdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-test-cmdline to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-test-compress-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-test-crypto-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-test-eventdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-test-fib to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-test-flow-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-test-gpudev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.677 Installing app/dpdk-test-pipeline to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.678 Installing app/dpdk-testpmd to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.678 Installing app/dpdk-test-regex to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.678 Installing app/dpdk-test-sad to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.678 Installing app/dpdk-test-security-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/config/rte_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/kvargs/rte_kvargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/telemetry/rte_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rtm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_alarm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitmap.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_branch_prediction.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bus.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_class.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_compat.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_debug.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_dev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_devargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_memconfig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_errno.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_epoll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_fbarray.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hexdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hypervisor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_interrupts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_keepalive.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_launch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_log.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_malloc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_mcslock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memory.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memzone.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_feature_defs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_features.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_per_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pflock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_random.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_reciprocal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqcount.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service_component.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_string_fns.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_tailq.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_thread.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_ticketlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_time.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point_register.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_uuid.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_version.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_vfio.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/linux/include/rte_os.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_c11_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_generic_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_zc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/rcu/rte_rcu_qsbr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.678 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_ptype.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_pool_ops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_dyn.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ip.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_udp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_esp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_sctp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_icmp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_arp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ether.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_macsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_vxlan.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gre.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gtp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_mpls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_higig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ecpri.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_geneve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_l2tpv2.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ppp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/meter/rte_meter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_cman.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_dev_info.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_eth_ctrl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/pci/rte_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_num.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_ipaddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_etheraddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_string.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_rdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_vt100.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_socket.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_cirbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_portlist.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_fbk_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_jhash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_sw.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_x86_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/timer/rte_timer.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl_osdep.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_op.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/bitratestats/rte_bitrate.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/bpf_def.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cfgfile/rte_cfgfile.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_compressdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_comp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_sym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_asym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/distributor/rte_distributor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/efd/rte_efd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_crypto_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_rx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_tx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_timer_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/gpudev/rte_gpudev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/gro/rte_gro.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/gso/rte_gso.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/ip_frag/rte_ip_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/jobstats/rte_jobstats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/latencystats/rte_latencystats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_scalar.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/member/rte_member.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/pcapng/rte_pcapng.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_empty_poll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_intel_uncore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_pmd_mgmt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_guest_channel.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.679 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/reorder/rte_reorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_approx.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_red.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_pie.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_std.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_c11.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_stubs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vdpa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_async.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sad.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ras.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sym_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/pdump/rte_pdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_em.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_learner.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_selector.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_wm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_array.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_cuckoo.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm_ipv6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_stub.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_port_in_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_table_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_extern.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ctl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip4_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_eth_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/pci/rte_bus_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/vdev/rte_bus_vdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/drivers/net/i40e/rte_pmd_i40e.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-devbind.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-pmdinfo.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-telemetry.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-hugepages.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/rte_build_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk-libs.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:02:54.680 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:02:54.680 Installing symlink pointing to librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so.23 00:02:54.680 Installing symlink pointing to librte_kvargs.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so 00:02:54.680 Installing symlink pointing to librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so.23 00:02:54.680 Installing symlink pointing to librte_telemetry.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so 00:02:54.680 Installing symlink pointing to librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so.23 00:02:54.680 Installing symlink pointing to librte_eal.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so 00:02:54.680 Installing symlink pointing to librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so.23 00:02:54.680 Installing symlink pointing to librte_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so 00:02:54.680 Installing symlink pointing to librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so.23 00:02:54.680 Installing symlink pointing to librte_rcu.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so 00:02:54.680 Installing symlink pointing to librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so.23 00:02:54.680 Installing symlink pointing to librte_mempool.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so 00:02:54.680 Installing symlink pointing to librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so.23 00:02:54.680 Installing symlink pointing to librte_mbuf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so 00:02:54.680 Installing symlink pointing to librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so.23 00:02:54.680 Installing symlink pointing to librte_net.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so 00:02:54.680 Installing symlink pointing to librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so.23 00:02:54.680 Installing symlink pointing to librte_meter.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so 00:02:54.680 Installing symlink pointing to librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so.23 00:02:54.680 Installing symlink pointing to librte_ethdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so 00:02:54.680 Installing symlink pointing to librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so.23 00:02:54.680 Installing symlink pointing to librte_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so 00:02:54.680 Installing symlink pointing to librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so.23 00:02:54.680 Installing symlink pointing to librte_cmdline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so 00:02:54.680 Installing symlink pointing to librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so.23 00:02:54.680 Installing symlink pointing to librte_metrics.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so 00:02:54.680 Installing symlink pointing to librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so.23 00:02:54.680 Installing symlink pointing to librte_hash.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so 00:02:54.680 Installing symlink pointing to librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so.23 00:02:54.680 Installing symlink pointing to librte_timer.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so 00:02:54.680 Installing symlink pointing to librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so.23 00:02:54.680 Installing symlink pointing to librte_acl.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so 00:02:54.680 Installing symlink pointing to librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so.23 00:02:54.680 Installing symlink pointing to librte_bbdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so 00:02:54.680 Installing symlink pointing to librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so.23 00:02:54.680 Installing symlink pointing to librte_bitratestats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so 00:02:54.680 Installing symlink pointing to librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so.23 00:02:54.680 Installing symlink pointing to librte_bpf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so 00:02:54.680 Installing symlink pointing to librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so.23 00:02:54.680 Installing symlink pointing to librte_cfgfile.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so 00:02:54.680 Installing symlink pointing to librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so.23 00:02:54.680 Installing symlink pointing to librte_compressdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so 00:02:54.680 Installing symlink pointing to librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so.23 00:02:54.680 Installing symlink pointing to librte_cryptodev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so 00:02:54.680 Installing symlink pointing to librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so.23 00:02:54.680 Installing symlink pointing to librte_distributor.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so 00:02:54.680 Installing symlink pointing to librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so.23 00:02:54.680 Installing symlink pointing to librte_efd.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so 00:02:54.680 Installing symlink pointing to librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so.23 00:02:54.680 Installing symlink pointing to librte_eventdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so 00:02:54.680 Installing symlink pointing to librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so.23 00:02:54.680 './librte_bus_pci.so' -> 'dpdk/pmds-23.0/librte_bus_pci.so' 00:02:54.680 './librte_bus_pci.so.23' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23' 00:02:54.680 './librte_bus_pci.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23.0' 00:02:54.680 './librte_bus_vdev.so' -> 'dpdk/pmds-23.0/librte_bus_vdev.so' 00:02:54.680 './librte_bus_vdev.so.23' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23' 00:02:54.680 './librte_bus_vdev.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23.0' 00:02:54.680 './librte_mempool_ring.so' -> 'dpdk/pmds-23.0/librte_mempool_ring.so' 00:02:54.680 './librte_mempool_ring.so.23' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23' 00:02:54.680 './librte_mempool_ring.so.23.0' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23.0' 00:02:54.680 './librte_net_i40e.so' -> 'dpdk/pmds-23.0/librte_net_i40e.so' 00:02:54.680 './librte_net_i40e.so.23' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23' 00:02:54.680 './librte_net_i40e.so.23.0' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23.0' 00:02:54.680 Installing symlink pointing to librte_gpudev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so 00:02:54.680 Installing symlink pointing to librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so.23 00:02:54.680 Installing symlink pointing to librte_gro.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so 00:02:54.680 Installing symlink pointing to librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so.23 00:02:54.680 Installing symlink pointing to librte_gso.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so 00:02:54.680 Installing symlink pointing to librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so.23 00:02:54.680 Installing symlink pointing to librte_ip_frag.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so 00:02:54.680 Installing symlink pointing to librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so.23 00:02:54.680 Installing symlink pointing to librte_jobstats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so 00:02:54.680 Installing symlink pointing to librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so.23 00:02:54.680 Installing symlink pointing to librte_latencystats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so 00:02:54.680 Installing symlink pointing to librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so.23 00:02:54.680 Installing symlink pointing to librte_lpm.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so 00:02:54.680 Installing symlink pointing to librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so.23 00:02:54.680 Installing symlink pointing to librte_member.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so 00:02:54.680 Installing symlink pointing to librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so.23 00:02:54.681 Installing symlink pointing to librte_pcapng.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so 00:02:54.681 Installing symlink pointing to librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so.23 00:02:54.681 Installing symlink pointing to librte_power.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so 00:02:54.681 Installing symlink pointing to librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so.23 00:02:54.681 Installing symlink pointing to librte_rawdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so 00:02:54.681 Installing symlink pointing to librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so.23 00:02:54.681 Installing symlink pointing to librte_regexdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so 00:02:54.681 Installing symlink pointing to librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so.23 00:02:54.681 Installing symlink pointing to librte_dmadev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so 00:02:54.681 Installing symlink pointing to librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so.23 00:02:54.681 Installing symlink pointing to librte_rib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so 00:02:54.681 Installing symlink pointing to librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so.23 00:02:54.681 Installing symlink pointing to librte_reorder.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so 00:02:54.681 Installing symlink pointing to librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so.23 00:02:54.681 Installing symlink pointing to librte_sched.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so 00:02:54.681 Installing symlink pointing to librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so.23 00:02:54.681 Installing symlink pointing to librte_security.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so 00:02:54.681 Installing symlink pointing to librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so.23 00:02:54.681 Installing symlink pointing to librte_stack.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so 00:02:54.681 Installing symlink pointing to librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so.23 00:02:54.681 Installing symlink pointing to librte_vhost.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so 00:02:54.681 Installing symlink pointing to librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so.23 00:02:54.681 Installing symlink pointing to librte_ipsec.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so 00:02:54.681 Installing symlink pointing to librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so.23 00:02:54.681 Installing symlink pointing to librte_fib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so 00:02:54.681 Installing symlink pointing to librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so.23 00:02:54.681 Installing symlink pointing to librte_port.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so 00:02:54.681 Installing symlink pointing to librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so.23 00:02:54.681 Installing symlink pointing to librte_pdump.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so 00:02:54.681 Installing symlink pointing to librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so.23 00:02:54.681 Installing symlink pointing to librte_table.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so 00:02:54.681 Installing symlink pointing to librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so.23 00:02:54.681 Installing symlink pointing to librte_pipeline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so 00:02:54.681 Installing symlink pointing to librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so.23 00:02:54.681 Installing symlink pointing to librte_graph.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so 00:02:54.681 Installing symlink pointing to librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so.23 00:02:54.681 Installing symlink pointing to librte_node.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so 00:02:54.681 Installing symlink pointing to librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23 00:02:54.681 Installing symlink pointing to librte_bus_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:02:54.681 Installing symlink pointing to librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23 00:02:54.681 Installing symlink pointing to librte_bus_vdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:02:54.681 Installing symlink pointing to librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23 00:02:54.681 Installing symlink pointing to librte_mempool_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:02:54.681 Installing symlink pointing to librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23 00:02:54.681 Installing symlink pointing to librte_net_i40e.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:02:54.681 Running custom install script '/bin/sh /home/vagrant/spdk_repo/dpdk/config/../buildtools/symlink-drivers-solibs.sh lib dpdk/pmds-23.0' 00:02:54.941 05:30:28 build_native_dpdk -- common/autobuild_common.sh@220 -- $ cat 00:02:54.941 05:30:28 build_native_dpdk -- common/autobuild_common.sh@225 -- $ cd /home/vagrant/spdk_repo/spdk 00:02:54.941 00:02:54.941 real 0m42.724s 00:02:54.941 user 4m15.065s 00:02:54.941 sys 0m48.254s 00:02:54.941 05:30:28 build_native_dpdk -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:02:54.941 05:30:28 build_native_dpdk -- common/autotest_common.sh@10 -- $ set +x 00:02:54.941 ************************************ 00:02:54.941 END TEST build_native_dpdk 00:02:54.941 ************************************ 00:02:54.941 05:30:28 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:02:54.941 05:30:28 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:02:54.941 05:30:28 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:02:54.941 05:30:28 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:02:54.942 05:30:28 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:02:54.942 05:30:28 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:02:54.942 05:30:28 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:02:54.942 05:30:28 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build --with-shared 00:02:54.942 Using /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig for additional libs... 00:02:55.202 DPDK libraries: /home/vagrant/spdk_repo/dpdk/build/lib 00:02:55.202 DPDK includes: //home/vagrant/spdk_repo/dpdk/build/include 00:02:55.202 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:02:55.462 Using 'verbs' RDMA provider 00:03:11.731 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:03:26.612 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:03:27.182 Creating mk/config.mk...done. 00:03:27.182 Creating mk/cc.flags.mk...done. 00:03:27.182 Type 'make' to build. 00:03:27.182 05:31:00 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:03:27.182 05:31:00 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:03:27.182 05:31:00 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:03:27.182 05:31:00 -- common/autotest_common.sh@10 -- $ set +x 00:03:27.182 ************************************ 00:03:27.182 START TEST make 00:03:27.182 ************************************ 00:03:27.182 05:31:00 make -- common/autotest_common.sh@1129 -- $ make -j10 00:03:27.750 make[1]: Nothing to be done for 'all'. 00:04:14.432 CC lib/ut_mock/mock.o 00:04:14.432 CC lib/ut/ut.o 00:04:14.432 CC lib/log/log_flags.o 00:04:14.432 CC lib/log/log.o 00:04:14.432 CC lib/log/log_deprecated.o 00:04:14.432 LIB libspdk_ut_mock.a 00:04:14.432 LIB libspdk_ut.a 00:04:14.432 LIB libspdk_log.a 00:04:14.433 SO libspdk_ut.so.2.0 00:04:14.433 SO libspdk_ut_mock.so.6.0 00:04:14.433 SO libspdk_log.so.7.1 00:04:14.433 SYMLINK libspdk_ut.so 00:04:14.433 SYMLINK libspdk_ut_mock.so 00:04:14.433 SYMLINK libspdk_log.so 00:04:14.433 CC lib/util/cpuset.o 00:04:14.433 CC lib/util/bit_array.o 00:04:14.433 CC lib/util/base64.o 00:04:14.433 CC lib/util/crc16.o 00:04:14.433 CC lib/util/crc32c.o 00:04:14.433 CC lib/util/crc32.o 00:04:14.433 CC lib/dma/dma.o 00:04:14.433 CXX lib/trace_parser/trace.o 00:04:14.433 CC lib/ioat/ioat.o 00:04:14.433 CC lib/vfio_user/host/vfio_user_pci.o 00:04:14.433 CC lib/util/crc32_ieee.o 00:04:14.433 CC lib/util/crc64.o 00:04:14.433 CC lib/util/dif.o 00:04:14.433 CC lib/util/fd.o 00:04:14.433 LIB libspdk_dma.a 00:04:14.433 CC lib/util/fd_group.o 00:04:14.433 CC lib/util/file.o 00:04:14.433 SO libspdk_dma.so.5.0 00:04:14.433 CC lib/util/hexlify.o 00:04:14.433 CC lib/vfio_user/host/vfio_user.o 00:04:14.433 SYMLINK libspdk_dma.so 00:04:14.433 CC lib/util/iov.o 00:04:14.433 CC lib/util/math.o 00:04:14.433 LIB libspdk_ioat.a 00:04:14.433 SO libspdk_ioat.so.7.0 00:04:14.433 CC lib/util/net.o 00:04:14.433 CC lib/util/pipe.o 00:04:14.433 CC lib/util/strerror_tls.o 00:04:14.433 SYMLINK libspdk_ioat.so 00:04:14.433 CC lib/util/string.o 00:04:14.433 CC lib/util/uuid.o 00:04:14.433 LIB libspdk_vfio_user.a 00:04:14.433 CC lib/util/xor.o 00:04:14.433 SO libspdk_vfio_user.so.5.0 00:04:14.433 CC lib/util/zipf.o 00:04:14.433 CC lib/util/md5.o 00:04:14.433 SYMLINK libspdk_vfio_user.so 00:04:14.433 LIB libspdk_util.a 00:04:14.433 SO libspdk_util.so.10.1 00:04:14.433 LIB libspdk_trace_parser.a 00:04:14.433 SYMLINK libspdk_util.so 00:04:14.433 SO libspdk_trace_parser.so.6.0 00:04:14.433 SYMLINK libspdk_trace_parser.so 00:04:14.433 CC lib/vmd/vmd.o 00:04:14.433 CC lib/vmd/led.o 00:04:14.433 CC lib/idxd/idxd.o 00:04:14.433 CC lib/idxd/idxd_user.o 00:04:14.433 CC lib/idxd/idxd_kernel.o 00:04:14.433 CC lib/rdma_utils/rdma_utils.o 00:04:14.433 CC lib/conf/conf.o 00:04:14.433 CC lib/env_dpdk/env.o 00:04:14.433 CC lib/env_dpdk/memory.o 00:04:14.433 CC lib/json/json_parse.o 00:04:14.433 CC lib/json/json_util.o 00:04:14.433 CC lib/json/json_write.o 00:04:14.433 LIB libspdk_conf.a 00:04:14.433 CC lib/env_dpdk/pci.o 00:04:14.433 SO libspdk_conf.so.6.0 00:04:14.433 CC lib/env_dpdk/init.o 00:04:14.433 LIB libspdk_rdma_utils.a 00:04:14.433 SYMLINK libspdk_conf.so 00:04:14.433 SO libspdk_rdma_utils.so.1.0 00:04:14.433 CC lib/env_dpdk/threads.o 00:04:14.433 SYMLINK libspdk_rdma_utils.so 00:04:14.433 CC lib/env_dpdk/pci_ioat.o 00:04:14.433 CC lib/env_dpdk/pci_virtio.o 00:04:14.433 CC lib/env_dpdk/pci_vmd.o 00:04:14.433 LIB libspdk_json.a 00:04:14.433 CC lib/env_dpdk/pci_idxd.o 00:04:14.433 SO libspdk_json.so.6.0 00:04:14.433 CC lib/env_dpdk/pci_event.o 00:04:14.433 CC lib/rdma_provider/common.o 00:04:14.433 SYMLINK libspdk_json.so 00:04:14.433 CC lib/rdma_provider/rdma_provider_verbs.o 00:04:14.433 CC lib/env_dpdk/sigbus_handler.o 00:04:14.433 CC lib/env_dpdk/pci_dpdk.o 00:04:14.433 CC lib/env_dpdk/pci_dpdk_2207.o 00:04:14.433 CC lib/env_dpdk/pci_dpdk_2211.o 00:04:14.433 LIB libspdk_vmd.a 00:04:14.433 LIB libspdk_idxd.a 00:04:14.433 SO libspdk_vmd.so.6.0 00:04:14.433 SO libspdk_idxd.so.12.1 00:04:14.433 LIB libspdk_rdma_provider.a 00:04:14.433 SYMLINK libspdk_vmd.so 00:04:14.433 CC lib/jsonrpc/jsonrpc_client.o 00:04:14.433 CC lib/jsonrpc/jsonrpc_server.o 00:04:14.433 SYMLINK libspdk_idxd.so 00:04:14.433 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:04:14.433 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:04:14.433 SO libspdk_rdma_provider.so.7.0 00:04:14.433 SYMLINK libspdk_rdma_provider.so 00:04:14.433 LIB libspdk_jsonrpc.a 00:04:14.433 SO libspdk_jsonrpc.so.6.0 00:04:14.433 SYMLINK libspdk_jsonrpc.so 00:04:14.433 CC lib/rpc/rpc.o 00:04:14.433 LIB libspdk_env_dpdk.a 00:04:14.433 SO libspdk_env_dpdk.so.15.1 00:04:14.433 LIB libspdk_rpc.a 00:04:14.433 SYMLINK libspdk_env_dpdk.so 00:04:14.433 SO libspdk_rpc.so.6.0 00:04:14.433 SYMLINK libspdk_rpc.so 00:04:14.433 CC lib/trace/trace.o 00:04:14.433 CC lib/trace/trace_flags.o 00:04:14.433 CC lib/trace/trace_rpc.o 00:04:14.433 CC lib/keyring/keyring.o 00:04:14.433 CC lib/keyring/keyring_rpc.o 00:04:14.433 CC lib/notify/notify_rpc.o 00:04:14.433 CC lib/notify/notify.o 00:04:14.433 LIB libspdk_notify.a 00:04:14.433 SO libspdk_notify.so.6.0 00:04:14.433 LIB libspdk_trace.a 00:04:14.433 SYMLINK libspdk_notify.so 00:04:14.433 LIB libspdk_keyring.a 00:04:14.433 SO libspdk_trace.so.11.0 00:04:14.433 SO libspdk_keyring.so.2.0 00:04:14.433 SYMLINK libspdk_keyring.so 00:04:14.433 SYMLINK libspdk_trace.so 00:04:14.433 CC lib/thread/thread.o 00:04:14.433 CC lib/thread/iobuf.o 00:04:14.433 CC lib/sock/sock.o 00:04:14.433 CC lib/sock/sock_rpc.o 00:04:14.433 LIB libspdk_sock.a 00:04:14.433 SO libspdk_sock.so.10.0 00:04:14.433 SYMLINK libspdk_sock.so 00:04:14.433 CC lib/nvme/nvme_ctrlr_cmd.o 00:04:14.433 CC lib/nvme/nvme_ctrlr.o 00:04:14.433 CC lib/nvme/nvme_fabric.o 00:04:14.433 CC lib/nvme/nvme_ns_cmd.o 00:04:14.433 CC lib/nvme/nvme_ns.o 00:04:14.433 CC lib/nvme/nvme_pcie_common.o 00:04:14.433 CC lib/nvme/nvme_pcie.o 00:04:14.433 CC lib/nvme/nvme_qpair.o 00:04:14.433 CC lib/nvme/nvme.o 00:04:14.433 LIB libspdk_thread.a 00:04:14.433 SO libspdk_thread.so.11.0 00:04:14.433 CC lib/nvme/nvme_quirks.o 00:04:14.433 SYMLINK libspdk_thread.so 00:04:14.433 CC lib/nvme/nvme_transport.o 00:04:14.433 CC lib/nvme/nvme_discovery.o 00:04:14.433 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:04:14.433 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:04:14.692 CC lib/nvme/nvme_tcp.o 00:04:14.692 CC lib/nvme/nvme_opal.o 00:04:14.692 CC lib/nvme/nvme_io_msg.o 00:04:14.692 CC lib/nvme/nvme_poll_group.o 00:04:14.951 CC lib/nvme/nvme_zns.o 00:04:14.951 CC lib/nvme/nvme_stubs.o 00:04:14.951 CC lib/nvme/nvme_auth.o 00:04:14.951 CC lib/nvme/nvme_cuse.o 00:04:15.209 CC lib/nvme/nvme_rdma.o 00:04:15.468 CC lib/accel/accel.o 00:04:15.468 CC lib/blob/blobstore.o 00:04:15.468 CC lib/init/json_config.o 00:04:15.468 CC lib/virtio/virtio.o 00:04:15.728 CC lib/fsdev/fsdev.o 00:04:15.728 CC lib/init/subsystem.o 00:04:15.986 CC lib/fsdev/fsdev_io.o 00:04:15.986 CC lib/virtio/virtio_vhost_user.o 00:04:15.986 CC lib/init/subsystem_rpc.o 00:04:15.986 CC lib/virtio/virtio_vfio_user.o 00:04:15.986 CC lib/virtio/virtio_pci.o 00:04:16.246 CC lib/blob/request.o 00:04:16.246 CC lib/init/rpc.o 00:04:16.246 CC lib/blob/zeroes.o 00:04:16.246 CC lib/accel/accel_rpc.o 00:04:16.246 LIB libspdk_init.a 00:04:16.246 CC lib/accel/accel_sw.o 00:04:16.246 CC lib/blob/blob_bs_dev.o 00:04:16.505 SO libspdk_init.so.6.0 00:04:16.505 CC lib/fsdev/fsdev_rpc.o 00:04:16.505 LIB libspdk_virtio.a 00:04:16.505 SYMLINK libspdk_init.so 00:04:16.505 SO libspdk_virtio.so.7.0 00:04:16.505 SYMLINK libspdk_virtio.so 00:04:16.505 LIB libspdk_fsdev.a 00:04:16.505 SO libspdk_fsdev.so.2.0 00:04:16.764 CC lib/event/app.o 00:04:16.764 CC lib/event/reactor.o 00:04:16.764 CC lib/event/scheduler_static.o 00:04:16.764 CC lib/event/log_rpc.o 00:04:16.764 CC lib/event/app_rpc.o 00:04:16.764 SYMLINK libspdk_fsdev.so 00:04:16.764 LIB libspdk_nvme.a 00:04:16.764 LIB libspdk_accel.a 00:04:16.764 SO libspdk_accel.so.16.0 00:04:16.764 SYMLINK libspdk_accel.so 00:04:17.022 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:04:17.022 SO libspdk_nvme.so.15.0 00:04:17.022 CC lib/bdev/bdev.o 00:04:17.022 CC lib/bdev/scsi_nvme.o 00:04:17.022 CC lib/bdev/bdev_rpc.o 00:04:17.022 CC lib/bdev/part.o 00:04:17.022 CC lib/bdev/bdev_zone.o 00:04:17.280 LIB libspdk_event.a 00:04:17.280 SYMLINK libspdk_nvme.so 00:04:17.280 SO libspdk_event.so.14.0 00:04:17.280 SYMLINK libspdk_event.so 00:04:17.539 LIB libspdk_fuse_dispatcher.a 00:04:17.539 SO libspdk_fuse_dispatcher.so.1.0 00:04:17.539 SYMLINK libspdk_fuse_dispatcher.so 00:04:18.977 LIB libspdk_blob.a 00:04:18.977 SO libspdk_blob.so.12.0 00:04:18.977 SYMLINK libspdk_blob.so 00:04:19.544 CC lib/blobfs/blobfs.o 00:04:19.544 CC lib/lvol/lvol.o 00:04:19.544 CC lib/blobfs/tree.o 00:04:20.111 LIB libspdk_bdev.a 00:04:20.111 SO libspdk_bdev.so.17.0 00:04:20.111 SYMLINK libspdk_bdev.so 00:04:20.369 LIB libspdk_blobfs.a 00:04:20.369 SO libspdk_blobfs.so.11.0 00:04:20.369 SYMLINK libspdk_blobfs.so 00:04:20.369 CC lib/nvmf/ctrlr_discovery.o 00:04:20.369 CC lib/nvmf/ctrlr.o 00:04:20.369 CC lib/nvmf/nvmf.o 00:04:20.369 CC lib/nvmf/ctrlr_bdev.o 00:04:20.369 CC lib/nvmf/subsystem.o 00:04:20.369 CC lib/scsi/dev.o 00:04:20.369 CC lib/ftl/ftl_core.o 00:04:20.369 CC lib/nbd/nbd.o 00:04:20.369 CC lib/ublk/ublk.o 00:04:20.628 LIB libspdk_lvol.a 00:04:20.628 SO libspdk_lvol.so.11.0 00:04:20.628 CC lib/scsi/lun.o 00:04:20.628 SYMLINK libspdk_lvol.so 00:04:20.628 CC lib/scsi/port.o 00:04:20.886 CC lib/ftl/ftl_init.o 00:04:20.886 CC lib/nbd/nbd_rpc.o 00:04:20.886 CC lib/ftl/ftl_layout.o 00:04:20.886 CC lib/ftl/ftl_debug.o 00:04:20.886 CC lib/scsi/scsi.o 00:04:20.886 CC lib/scsi/scsi_bdev.o 00:04:20.886 LIB libspdk_nbd.a 00:04:21.145 SO libspdk_nbd.so.7.0 00:04:21.145 CC lib/ftl/ftl_io.o 00:04:21.145 SYMLINK libspdk_nbd.so 00:04:21.145 CC lib/ftl/ftl_sb.o 00:04:21.145 CC lib/ftl/ftl_l2p.o 00:04:21.145 CC lib/ublk/ublk_rpc.o 00:04:21.145 CC lib/ftl/ftl_l2p_flat.o 00:04:21.145 CC lib/nvmf/nvmf_rpc.o 00:04:21.403 CC lib/ftl/ftl_nv_cache.o 00:04:21.403 LIB libspdk_ublk.a 00:04:21.403 CC lib/nvmf/transport.o 00:04:21.403 SO libspdk_ublk.so.3.0 00:04:21.403 SYMLINK libspdk_ublk.so 00:04:21.403 CC lib/scsi/scsi_pr.o 00:04:21.403 CC lib/scsi/scsi_rpc.o 00:04:21.403 CC lib/nvmf/tcp.o 00:04:21.403 CC lib/scsi/task.o 00:04:21.662 CC lib/ftl/ftl_band.o 00:04:21.662 CC lib/nvmf/stubs.o 00:04:21.662 CC lib/ftl/ftl_band_ops.o 00:04:21.662 LIB libspdk_scsi.a 00:04:21.919 CC lib/ftl/ftl_writer.o 00:04:21.919 SO libspdk_scsi.so.9.0 00:04:21.919 SYMLINK libspdk_scsi.so 00:04:21.919 CC lib/ftl/ftl_rq.o 00:04:22.176 CC lib/iscsi/conn.o 00:04:22.176 CC lib/nvmf/mdns_server.o 00:04:22.176 CC lib/nvmf/rdma.o 00:04:22.176 CC lib/ftl/ftl_reloc.o 00:04:22.176 CC lib/ftl/ftl_l2p_cache.o 00:04:22.176 CC lib/vhost/vhost.o 00:04:22.434 CC lib/ftl/ftl_p2l.o 00:04:22.434 CC lib/ftl/ftl_p2l_log.o 00:04:22.434 CC lib/ftl/mngt/ftl_mngt.o 00:04:22.434 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:04:22.434 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:04:22.691 CC lib/iscsi/init_grp.o 00:04:22.691 CC lib/vhost/vhost_rpc.o 00:04:22.691 CC lib/ftl/mngt/ftl_mngt_startup.o 00:04:22.691 CC lib/ftl/mngt/ftl_mngt_md.o 00:04:22.691 CC lib/ftl/mngt/ftl_mngt_misc.o 00:04:22.691 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:04:22.950 CC lib/iscsi/iscsi.o 00:04:22.950 CC lib/nvmf/auth.o 00:04:22.950 CC lib/vhost/vhost_scsi.o 00:04:22.950 CC lib/vhost/vhost_blk.o 00:04:22.950 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:04:22.950 CC lib/vhost/rte_vhost_user.o 00:04:23.209 CC lib/iscsi/param.o 00:04:23.209 CC lib/ftl/mngt/ftl_mngt_band.o 00:04:23.209 CC lib/iscsi/portal_grp.o 00:04:23.209 CC lib/iscsi/tgt_node.o 00:04:23.468 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:04:23.468 CC lib/iscsi/iscsi_subsystem.o 00:04:23.468 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:04:23.725 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:04:23.725 CC lib/iscsi/iscsi_rpc.o 00:04:23.725 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:04:23.725 CC lib/iscsi/task.o 00:04:23.725 CC lib/ftl/utils/ftl_conf.o 00:04:23.725 CC lib/ftl/utils/ftl_md.o 00:04:23.725 CC lib/ftl/utils/ftl_mempool.o 00:04:23.725 CC lib/ftl/utils/ftl_bitmap.o 00:04:23.983 CC lib/ftl/utils/ftl_property.o 00:04:23.983 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:04:23.983 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:04:23.983 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:04:23.983 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:04:23.983 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:04:23.983 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:04:24.240 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:04:24.240 CC lib/ftl/upgrade/ftl_sb_v3.o 00:04:24.240 CC lib/ftl/upgrade/ftl_sb_v5.o 00:04:24.240 LIB libspdk_vhost.a 00:04:24.240 CC lib/ftl/nvc/ftl_nvc_dev.o 00:04:24.241 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:04:24.241 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:04:24.241 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:04:24.241 SO libspdk_vhost.so.8.0 00:04:24.241 CC lib/ftl/base/ftl_base_dev.o 00:04:24.497 CC lib/ftl/base/ftl_base_bdev.o 00:04:24.497 SYMLINK libspdk_vhost.so 00:04:24.497 CC lib/ftl/ftl_trace.o 00:04:24.497 LIB libspdk_iscsi.a 00:04:24.755 LIB libspdk_ftl.a 00:04:24.755 SO libspdk_iscsi.so.8.0 00:04:24.755 LIB libspdk_nvmf.a 00:04:24.755 SO libspdk_ftl.so.9.0 00:04:25.013 SYMLINK libspdk_iscsi.so 00:04:25.013 SO libspdk_nvmf.so.20.0 00:04:25.271 SYMLINK libspdk_ftl.so 00:04:25.271 SYMLINK libspdk_nvmf.so 00:04:25.528 CC module/env_dpdk/env_dpdk_rpc.o 00:04:25.528 CC module/fsdev/aio/fsdev_aio.o 00:04:25.528 CC module/sock/posix/posix.o 00:04:25.528 CC module/scheduler/gscheduler/gscheduler.o 00:04:25.528 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:04:25.528 CC module/accel/error/accel_error.o 00:04:25.528 CC module/keyring/linux/keyring.o 00:04:25.528 CC module/scheduler/dynamic/scheduler_dynamic.o 00:04:25.528 CC module/keyring/file/keyring.o 00:04:25.784 CC module/blob/bdev/blob_bdev.o 00:04:25.784 LIB libspdk_env_dpdk_rpc.a 00:04:25.784 SO libspdk_env_dpdk_rpc.so.6.0 00:04:25.784 SYMLINK libspdk_env_dpdk_rpc.so 00:04:25.784 CC module/keyring/linux/keyring_rpc.o 00:04:25.784 LIB libspdk_scheduler_gscheduler.a 00:04:25.784 CC module/keyring/file/keyring_rpc.o 00:04:25.784 CC module/accel/error/accel_error_rpc.o 00:04:25.784 SO libspdk_scheduler_gscheduler.so.4.0 00:04:25.784 LIB libspdk_scheduler_dpdk_governor.a 00:04:25.784 SO libspdk_scheduler_dpdk_governor.so.4.0 00:04:25.784 SYMLINK libspdk_scheduler_gscheduler.so 00:04:25.784 CC module/fsdev/aio/fsdev_aio_rpc.o 00:04:25.784 LIB libspdk_scheduler_dynamic.a 00:04:25.784 SYMLINK libspdk_scheduler_dpdk_governor.so 00:04:25.784 SO libspdk_scheduler_dynamic.so.4.0 00:04:25.784 CC module/fsdev/aio/linux_aio_mgr.o 00:04:25.784 LIB libspdk_keyring_linux.a 00:04:26.042 LIB libspdk_keyring_file.a 00:04:26.042 LIB libspdk_accel_error.a 00:04:26.042 SO libspdk_keyring_linux.so.1.0 00:04:26.042 LIB libspdk_blob_bdev.a 00:04:26.042 SO libspdk_keyring_file.so.2.0 00:04:26.042 SYMLINK libspdk_scheduler_dynamic.so 00:04:26.042 SO libspdk_accel_error.so.2.0 00:04:26.042 SO libspdk_blob_bdev.so.12.0 00:04:26.042 SYMLINK libspdk_keyring_linux.so 00:04:26.042 SYMLINK libspdk_keyring_file.so 00:04:26.042 SYMLINK libspdk_accel_error.so 00:04:26.042 SYMLINK libspdk_blob_bdev.so 00:04:26.042 CC module/accel/ioat/accel_ioat.o 00:04:26.042 CC module/accel/ioat/accel_ioat_rpc.o 00:04:26.042 CC module/accel/dsa/accel_dsa.o 00:04:26.300 CC module/accel/iaa/accel_iaa.o 00:04:26.300 LIB libspdk_accel_ioat.a 00:04:26.300 CC module/bdev/delay/vbdev_delay.o 00:04:26.300 CC module/blobfs/bdev/blobfs_bdev.o 00:04:26.300 CC module/bdev/gpt/gpt.o 00:04:26.300 CC module/bdev/error/vbdev_error.o 00:04:26.300 SO libspdk_accel_ioat.so.6.0 00:04:26.300 CC module/bdev/lvol/vbdev_lvol.o 00:04:26.300 SYMLINK libspdk_accel_ioat.so 00:04:26.300 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:04:26.300 CC module/accel/iaa/accel_iaa_rpc.o 00:04:26.300 CC module/accel/dsa/accel_dsa_rpc.o 00:04:26.300 LIB libspdk_sock_posix.a 00:04:26.558 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:04:26.558 CC module/bdev/gpt/vbdev_gpt.o 00:04:26.558 SO libspdk_sock_posix.so.6.0 00:04:26.558 LIB libspdk_fsdev_aio.a 00:04:26.558 SO libspdk_fsdev_aio.so.1.0 00:04:26.558 SYMLINK libspdk_sock_posix.so 00:04:26.558 LIB libspdk_accel_iaa.a 00:04:26.558 CC module/bdev/error/vbdev_error_rpc.o 00:04:26.558 LIB libspdk_accel_dsa.a 00:04:26.558 SO libspdk_accel_iaa.so.3.0 00:04:26.558 SO libspdk_accel_dsa.so.5.0 00:04:26.558 LIB libspdk_blobfs_bdev.a 00:04:26.558 SYMLINK libspdk_fsdev_aio.so 00:04:26.558 CC module/bdev/delay/vbdev_delay_rpc.o 00:04:26.558 SO libspdk_blobfs_bdev.so.6.0 00:04:26.558 SYMLINK libspdk_accel_dsa.so 00:04:26.558 SYMLINK libspdk_accel_iaa.so 00:04:26.816 SYMLINK libspdk_blobfs_bdev.so 00:04:26.816 LIB libspdk_bdev_error.a 00:04:26.816 LIB libspdk_bdev_gpt.a 00:04:26.816 SO libspdk_bdev_error.so.6.0 00:04:26.816 SO libspdk_bdev_gpt.so.6.0 00:04:26.816 CC module/bdev/malloc/bdev_malloc.o 00:04:26.816 CC module/bdev/null/bdev_null.o 00:04:26.816 CC module/bdev/passthru/vbdev_passthru.o 00:04:26.816 CC module/bdev/nvme/bdev_nvme.o 00:04:26.816 SYMLINK libspdk_bdev_gpt.so 00:04:26.816 SYMLINK libspdk_bdev_error.so 00:04:26.816 CC module/bdev/nvme/bdev_nvme_rpc.o 00:04:26.816 CC module/bdev/nvme/nvme_rpc.o 00:04:26.816 LIB libspdk_bdev_delay.a 00:04:26.816 CC module/bdev/raid/bdev_raid.o 00:04:26.816 CC module/bdev/split/vbdev_split.o 00:04:26.816 SO libspdk_bdev_delay.so.6.0 00:04:26.816 LIB libspdk_bdev_lvol.a 00:04:26.816 SO libspdk_bdev_lvol.so.6.0 00:04:26.816 SYMLINK libspdk_bdev_delay.so 00:04:26.816 CC module/bdev/raid/bdev_raid_rpc.o 00:04:27.074 SYMLINK libspdk_bdev_lvol.so 00:04:27.074 CC module/bdev/null/bdev_null_rpc.o 00:04:27.074 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:04:27.074 LIB libspdk_bdev_null.a 00:04:27.074 CC module/bdev/split/vbdev_split_rpc.o 00:04:27.074 CC module/bdev/nvme/bdev_mdns_client.o 00:04:27.074 SO libspdk_bdev_null.so.6.0 00:04:27.074 CC module/bdev/malloc/bdev_malloc_rpc.o 00:04:27.074 LIB libspdk_bdev_passthru.a 00:04:27.074 SYMLINK libspdk_bdev_null.so 00:04:27.331 SO libspdk_bdev_passthru.so.6.0 00:04:27.331 CC module/bdev/zone_block/vbdev_zone_block.o 00:04:27.331 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:04:27.331 SYMLINK libspdk_bdev_passthru.so 00:04:27.331 CC module/bdev/nvme/vbdev_opal.o 00:04:27.331 CC module/bdev/nvme/vbdev_opal_rpc.o 00:04:27.331 LIB libspdk_bdev_split.a 00:04:27.331 LIB libspdk_bdev_malloc.a 00:04:27.331 SO libspdk_bdev_split.so.6.0 00:04:27.331 CC module/bdev/aio/bdev_aio.o 00:04:27.331 SYMLINK libspdk_bdev_split.so 00:04:27.331 SO libspdk_bdev_malloc.so.6.0 00:04:27.331 CC module/bdev/raid/bdev_raid_sb.o 00:04:27.331 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:04:27.331 SYMLINK libspdk_bdev_malloc.so 00:04:27.588 CC module/bdev/aio/bdev_aio_rpc.o 00:04:27.588 LIB libspdk_bdev_zone_block.a 00:04:27.588 CC module/bdev/ftl/bdev_ftl.o 00:04:27.588 SO libspdk_bdev_zone_block.so.6.0 00:04:27.588 CC module/bdev/raid/raid0.o 00:04:27.588 CC module/bdev/iscsi/bdev_iscsi.o 00:04:27.588 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:04:27.588 SYMLINK libspdk_bdev_zone_block.so 00:04:27.588 CC module/bdev/raid/raid1.o 00:04:27.588 CC module/bdev/raid/concat.o 00:04:27.588 CC module/bdev/virtio/bdev_virtio_scsi.o 00:04:27.846 LIB libspdk_bdev_aio.a 00:04:27.846 CC module/bdev/virtio/bdev_virtio_blk.o 00:04:27.846 SO libspdk_bdev_aio.so.6.0 00:04:27.846 CC module/bdev/virtio/bdev_virtio_rpc.o 00:04:27.846 CC module/bdev/ftl/bdev_ftl_rpc.o 00:04:27.846 SYMLINK libspdk_bdev_aio.so 00:04:27.846 CC module/bdev/raid/raid5f.o 00:04:28.105 LIB libspdk_bdev_iscsi.a 00:04:28.105 SO libspdk_bdev_iscsi.so.6.0 00:04:28.105 LIB libspdk_bdev_ftl.a 00:04:28.105 SYMLINK libspdk_bdev_iscsi.so 00:04:28.105 SO libspdk_bdev_ftl.so.6.0 00:04:28.105 SYMLINK libspdk_bdev_ftl.so 00:04:28.363 LIB libspdk_bdev_virtio.a 00:04:28.363 SO libspdk_bdev_virtio.so.6.0 00:04:28.363 SYMLINK libspdk_bdev_virtio.so 00:04:28.363 LIB libspdk_bdev_raid.a 00:04:28.363 SO libspdk_bdev_raid.so.6.0 00:04:28.621 SYMLINK libspdk_bdev_raid.so 00:04:29.557 LIB libspdk_bdev_nvme.a 00:04:29.557 SO libspdk_bdev_nvme.so.7.1 00:04:29.557 SYMLINK libspdk_bdev_nvme.so 00:04:30.121 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:04:30.121 CC module/event/subsystems/sock/sock.o 00:04:30.380 CC module/event/subsystems/vmd/vmd.o 00:04:30.380 CC module/event/subsystems/vmd/vmd_rpc.o 00:04:30.380 CC module/event/subsystems/scheduler/scheduler.o 00:04:30.380 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:04:30.380 CC module/event/subsystems/iobuf/iobuf.o 00:04:30.380 CC module/event/subsystems/fsdev/fsdev.o 00:04:30.380 CC module/event/subsystems/keyring/keyring.o 00:04:30.380 LIB libspdk_event_keyring.a 00:04:30.380 LIB libspdk_event_sock.a 00:04:30.380 LIB libspdk_event_fsdev.a 00:04:30.380 LIB libspdk_event_vhost_blk.a 00:04:30.380 LIB libspdk_event_vmd.a 00:04:30.380 LIB libspdk_event_iobuf.a 00:04:30.380 SO libspdk_event_keyring.so.1.0 00:04:30.380 SO libspdk_event_sock.so.5.0 00:04:30.380 SO libspdk_event_fsdev.so.1.0 00:04:30.380 LIB libspdk_event_scheduler.a 00:04:30.380 SO libspdk_event_vhost_blk.so.3.0 00:04:30.380 SO libspdk_event_vmd.so.6.0 00:04:30.380 SO libspdk_event_iobuf.so.3.0 00:04:30.380 SO libspdk_event_scheduler.so.4.0 00:04:30.380 SYMLINK libspdk_event_keyring.so 00:04:30.380 SYMLINK libspdk_event_sock.so 00:04:30.380 SYMLINK libspdk_event_fsdev.so 00:04:30.380 SYMLINK libspdk_event_vhost_blk.so 00:04:30.380 SYMLINK libspdk_event_vmd.so 00:04:30.380 SYMLINK libspdk_event_iobuf.so 00:04:30.380 SYMLINK libspdk_event_scheduler.so 00:04:30.947 CC module/event/subsystems/accel/accel.o 00:04:30.947 LIB libspdk_event_accel.a 00:04:30.947 SO libspdk_event_accel.so.6.0 00:04:31.205 SYMLINK libspdk_event_accel.so 00:04:31.465 CC module/event/subsystems/bdev/bdev.o 00:04:31.724 LIB libspdk_event_bdev.a 00:04:31.724 SO libspdk_event_bdev.so.6.0 00:04:31.724 SYMLINK libspdk_event_bdev.so 00:04:32.293 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:04:32.293 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:04:32.293 CC module/event/subsystems/scsi/scsi.o 00:04:32.293 CC module/event/subsystems/nbd/nbd.o 00:04:32.293 CC module/event/subsystems/ublk/ublk.o 00:04:32.293 LIB libspdk_event_nbd.a 00:04:32.293 LIB libspdk_event_scsi.a 00:04:32.293 LIB libspdk_event_ublk.a 00:04:32.293 SO libspdk_event_nbd.so.6.0 00:04:32.293 SO libspdk_event_scsi.so.6.0 00:04:32.293 SO libspdk_event_ublk.so.3.0 00:04:32.293 SYMLINK libspdk_event_nbd.so 00:04:32.293 SYMLINK libspdk_event_scsi.so 00:04:32.293 LIB libspdk_event_nvmf.a 00:04:32.293 SYMLINK libspdk_event_ublk.so 00:04:32.551 SO libspdk_event_nvmf.so.6.0 00:04:32.551 SYMLINK libspdk_event_nvmf.so 00:04:32.809 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:04:32.809 CC module/event/subsystems/iscsi/iscsi.o 00:04:32.809 LIB libspdk_event_vhost_scsi.a 00:04:32.809 LIB libspdk_event_iscsi.a 00:04:32.809 SO libspdk_event_vhost_scsi.so.3.0 00:04:33.066 SO libspdk_event_iscsi.so.6.0 00:04:33.066 SYMLINK libspdk_event_vhost_scsi.so 00:04:33.066 SYMLINK libspdk_event_iscsi.so 00:04:33.324 SO libspdk.so.6.0 00:04:33.324 SYMLINK libspdk.so 00:04:33.584 TEST_HEADER include/spdk/accel.h 00:04:33.584 TEST_HEADER include/spdk/accel_module.h 00:04:33.584 CXX app/trace/trace.o 00:04:33.584 TEST_HEADER include/spdk/assert.h 00:04:33.584 TEST_HEADER include/spdk/barrier.h 00:04:33.584 CC app/trace_record/trace_record.o 00:04:33.584 TEST_HEADER include/spdk/base64.h 00:04:33.584 TEST_HEADER include/spdk/bdev.h 00:04:33.584 TEST_HEADER include/spdk/bdev_module.h 00:04:33.584 TEST_HEADER include/spdk/bdev_zone.h 00:04:33.584 TEST_HEADER include/spdk/bit_array.h 00:04:33.584 TEST_HEADER include/spdk/bit_pool.h 00:04:33.584 TEST_HEADER include/spdk/blob_bdev.h 00:04:33.584 TEST_HEADER include/spdk/blobfs_bdev.h 00:04:33.584 TEST_HEADER include/spdk/blobfs.h 00:04:33.584 TEST_HEADER include/spdk/blob.h 00:04:33.584 TEST_HEADER include/spdk/conf.h 00:04:33.584 TEST_HEADER include/spdk/config.h 00:04:33.584 TEST_HEADER include/spdk/cpuset.h 00:04:33.584 CC examples/interrupt_tgt/interrupt_tgt.o 00:04:33.584 TEST_HEADER include/spdk/crc16.h 00:04:33.584 TEST_HEADER include/spdk/crc32.h 00:04:33.584 TEST_HEADER include/spdk/crc64.h 00:04:33.584 TEST_HEADER include/spdk/dif.h 00:04:33.584 TEST_HEADER include/spdk/dma.h 00:04:33.584 TEST_HEADER include/spdk/endian.h 00:04:33.584 TEST_HEADER include/spdk/env_dpdk.h 00:04:33.584 TEST_HEADER include/spdk/env.h 00:04:33.584 TEST_HEADER include/spdk/event.h 00:04:33.584 TEST_HEADER include/spdk/fd_group.h 00:04:33.584 TEST_HEADER include/spdk/fd.h 00:04:33.584 TEST_HEADER include/spdk/file.h 00:04:33.584 TEST_HEADER include/spdk/fsdev.h 00:04:33.584 TEST_HEADER include/spdk/fsdev_module.h 00:04:33.584 TEST_HEADER include/spdk/ftl.h 00:04:33.584 TEST_HEADER include/spdk/fuse_dispatcher.h 00:04:33.584 TEST_HEADER include/spdk/gpt_spec.h 00:04:33.584 TEST_HEADER include/spdk/hexlify.h 00:04:33.584 TEST_HEADER include/spdk/histogram_data.h 00:04:33.584 TEST_HEADER include/spdk/idxd.h 00:04:33.584 CC examples/ioat/perf/perf.o 00:04:33.584 TEST_HEADER include/spdk/idxd_spec.h 00:04:33.584 TEST_HEADER include/spdk/init.h 00:04:33.584 TEST_HEADER include/spdk/ioat.h 00:04:33.584 CC examples/util/zipf/zipf.o 00:04:33.584 TEST_HEADER include/spdk/ioat_spec.h 00:04:33.584 TEST_HEADER include/spdk/iscsi_spec.h 00:04:33.584 CC test/thread/poller_perf/poller_perf.o 00:04:33.584 TEST_HEADER include/spdk/json.h 00:04:33.584 TEST_HEADER include/spdk/jsonrpc.h 00:04:33.584 TEST_HEADER include/spdk/keyring.h 00:04:33.584 TEST_HEADER include/spdk/keyring_module.h 00:04:33.584 TEST_HEADER include/spdk/likely.h 00:04:33.584 TEST_HEADER include/spdk/log.h 00:04:33.584 TEST_HEADER include/spdk/lvol.h 00:04:33.584 TEST_HEADER include/spdk/md5.h 00:04:33.584 TEST_HEADER include/spdk/memory.h 00:04:33.584 TEST_HEADER include/spdk/mmio.h 00:04:33.584 TEST_HEADER include/spdk/nbd.h 00:04:33.584 TEST_HEADER include/spdk/net.h 00:04:33.584 TEST_HEADER include/spdk/notify.h 00:04:33.584 TEST_HEADER include/spdk/nvme.h 00:04:33.584 CC test/dma/test_dma/test_dma.o 00:04:33.584 TEST_HEADER include/spdk/nvme_intel.h 00:04:33.584 TEST_HEADER include/spdk/nvme_ocssd.h 00:04:33.584 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:04:33.584 TEST_HEADER include/spdk/nvme_spec.h 00:04:33.584 TEST_HEADER include/spdk/nvme_zns.h 00:04:33.584 TEST_HEADER include/spdk/nvmf_cmd.h 00:04:33.584 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:04:33.584 TEST_HEADER include/spdk/nvmf.h 00:04:33.584 TEST_HEADER include/spdk/nvmf_spec.h 00:04:33.584 TEST_HEADER include/spdk/nvmf_transport.h 00:04:33.584 TEST_HEADER include/spdk/opal.h 00:04:33.584 TEST_HEADER include/spdk/opal_spec.h 00:04:33.584 TEST_HEADER include/spdk/pci_ids.h 00:04:33.584 CC test/app/bdev_svc/bdev_svc.o 00:04:33.584 TEST_HEADER include/spdk/pipe.h 00:04:33.584 TEST_HEADER include/spdk/queue.h 00:04:33.584 TEST_HEADER include/spdk/reduce.h 00:04:33.584 TEST_HEADER include/spdk/rpc.h 00:04:33.584 TEST_HEADER include/spdk/scheduler.h 00:04:33.584 TEST_HEADER include/spdk/scsi.h 00:04:33.584 TEST_HEADER include/spdk/scsi_spec.h 00:04:33.584 TEST_HEADER include/spdk/sock.h 00:04:33.584 TEST_HEADER include/spdk/stdinc.h 00:04:33.584 TEST_HEADER include/spdk/string.h 00:04:33.584 TEST_HEADER include/spdk/thread.h 00:04:33.584 CC test/env/mem_callbacks/mem_callbacks.o 00:04:33.584 TEST_HEADER include/spdk/trace.h 00:04:33.584 TEST_HEADER include/spdk/trace_parser.h 00:04:33.584 TEST_HEADER include/spdk/tree.h 00:04:33.584 TEST_HEADER include/spdk/ublk.h 00:04:33.584 TEST_HEADER include/spdk/util.h 00:04:33.584 TEST_HEADER include/spdk/uuid.h 00:04:33.584 TEST_HEADER include/spdk/version.h 00:04:33.584 TEST_HEADER include/spdk/vfio_user_pci.h 00:04:33.584 TEST_HEADER include/spdk/vfio_user_spec.h 00:04:33.584 TEST_HEADER include/spdk/vhost.h 00:04:33.584 TEST_HEADER include/spdk/vmd.h 00:04:33.584 TEST_HEADER include/spdk/xor.h 00:04:33.584 TEST_HEADER include/spdk/zipf.h 00:04:33.584 CXX test/cpp_headers/accel.o 00:04:33.844 LINK interrupt_tgt 00:04:33.844 LINK poller_perf 00:04:33.844 LINK zipf 00:04:33.844 LINK spdk_trace_record 00:04:33.844 LINK ioat_perf 00:04:33.844 LINK bdev_svc 00:04:33.844 CXX test/cpp_headers/accel_module.o 00:04:33.844 LINK mem_callbacks 00:04:33.844 CXX test/cpp_headers/assert.o 00:04:33.844 CXX test/cpp_headers/barrier.o 00:04:33.844 CXX test/cpp_headers/base64.o 00:04:33.844 LINK spdk_trace 00:04:34.104 CC test/env/vtophys/vtophys.o 00:04:34.104 CC examples/ioat/verify/verify.o 00:04:34.104 CXX test/cpp_headers/bdev.o 00:04:34.104 CXX test/cpp_headers/bdev_module.o 00:04:34.104 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:04:34.104 CC test/env/memory/memory_ut.o 00:04:34.104 CC test/env/pci/pci_ut.o 00:04:34.104 LINK test_dma 00:04:34.104 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:04:34.104 LINK vtophys 00:04:34.363 CC app/nvmf_tgt/nvmf_main.o 00:04:34.363 LINK env_dpdk_post_init 00:04:34.363 LINK verify 00:04:34.363 CXX test/cpp_headers/bdev_zone.o 00:04:34.363 CC test/rpc_client/rpc_client_test.o 00:04:34.363 CXX test/cpp_headers/bit_array.o 00:04:34.363 LINK nvmf_tgt 00:04:34.621 CXX test/cpp_headers/bit_pool.o 00:04:34.621 LINK rpc_client_test 00:04:34.621 CC test/event/event_perf/event_perf.o 00:04:34.621 CC test/event/reactor/reactor.o 00:04:34.621 LINK pci_ut 00:04:34.621 LINK nvme_fuzz 00:04:34.621 CC test/nvme/aer/aer.o 00:04:34.621 CC examples/thread/thread/thread_ex.o 00:04:34.621 CXX test/cpp_headers/blob_bdev.o 00:04:34.621 LINK reactor 00:04:34.621 LINK event_perf 00:04:34.880 CC app/iscsi_tgt/iscsi_tgt.o 00:04:34.880 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:04:34.880 CC app/spdk_tgt/spdk_tgt.o 00:04:34.880 CXX test/cpp_headers/blobfs_bdev.o 00:04:34.880 LINK thread 00:04:34.880 LINK aer 00:04:34.880 CC test/event/reactor_perf/reactor_perf.o 00:04:35.139 LINK iscsi_tgt 00:04:35.139 LINK memory_ut 00:04:35.139 CC test/accel/dif/dif.o 00:04:35.139 CC test/blobfs/mkfs/mkfs.o 00:04:35.139 CXX test/cpp_headers/blobfs.o 00:04:35.139 LINK reactor_perf 00:04:35.139 LINK spdk_tgt 00:04:35.139 CC test/nvme/reset/reset.o 00:04:35.139 CXX test/cpp_headers/blob.o 00:04:35.398 LINK mkfs 00:04:35.398 CC examples/sock/hello_world/hello_sock.o 00:04:35.398 CC app/spdk_lspci/spdk_lspci.o 00:04:35.398 CC test/event/app_repeat/app_repeat.o 00:04:35.398 CXX test/cpp_headers/conf.o 00:04:35.398 CC examples/vmd/lsvmd/lsvmd.o 00:04:35.398 LINK spdk_lspci 00:04:35.398 LINK reset 00:04:35.657 LINK app_repeat 00:04:35.657 LINK lsvmd 00:04:35.657 CC examples/vmd/led/led.o 00:04:35.657 LINK hello_sock 00:04:35.657 CXX test/cpp_headers/config.o 00:04:35.657 CC test/lvol/esnap/esnap.o 00:04:35.657 CXX test/cpp_headers/cpuset.o 00:04:35.657 LINK led 00:04:35.657 CC app/spdk_nvme_perf/perf.o 00:04:35.657 CC test/nvme/sgl/sgl.o 00:04:35.915 CXX test/cpp_headers/crc16.o 00:04:35.915 CC test/nvme/overhead/overhead.o 00:04:35.916 CC test/nvme/e2edp/nvme_dp.o 00:04:35.916 LINK dif 00:04:35.916 CC test/event/scheduler/scheduler.o 00:04:35.916 CXX test/cpp_headers/crc32.o 00:04:35.916 CC examples/idxd/perf/perf.o 00:04:36.173 CXX test/cpp_headers/crc64.o 00:04:36.173 LINK sgl 00:04:36.173 LINK scheduler 00:04:36.173 LINK nvme_dp 00:04:36.173 LINK overhead 00:04:36.173 CXX test/cpp_headers/dif.o 00:04:36.173 CC test/bdev/bdevio/bdevio.o 00:04:36.432 CC test/nvme/err_injection/err_injection.o 00:04:36.432 CXX test/cpp_headers/dma.o 00:04:36.432 CC test/nvme/startup/startup.o 00:04:36.432 LINK idxd_perf 00:04:36.432 CC examples/fsdev/hello_world/hello_fsdev.o 00:04:36.432 CC examples/accel/perf/accel_perf.o 00:04:36.432 CXX test/cpp_headers/endian.o 00:04:36.432 LINK startup 00:04:36.691 LINK err_injection 00:04:36.691 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:04:36.691 CXX test/cpp_headers/env_dpdk.o 00:04:36.691 LINK hello_fsdev 00:04:36.691 LINK spdk_nvme_perf 00:04:36.691 LINK bdevio 00:04:36.691 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:04:36.691 CC test/nvme/reserve/reserve.o 00:04:36.949 CC test/nvme/simple_copy/simple_copy.o 00:04:36.949 CXX test/cpp_headers/env.o 00:04:36.949 LINK iscsi_fuzz 00:04:36.949 CC app/spdk_nvme_identify/identify.o 00:04:36.949 CC app/spdk_top/spdk_top.o 00:04:36.949 CC app/spdk_nvme_discover/discovery_aer.o 00:04:36.949 CXX test/cpp_headers/event.o 00:04:36.949 LINK accel_perf 00:04:36.949 LINK reserve 00:04:37.208 LINK simple_copy 00:04:37.209 CC test/nvme/connect_stress/connect_stress.o 00:04:37.209 LINK vhost_fuzz 00:04:37.209 CXX test/cpp_headers/fd_group.o 00:04:37.209 LINK spdk_nvme_discover 00:04:37.209 CXX test/cpp_headers/fd.o 00:04:37.481 CC examples/nvme/hello_world/hello_world.o 00:04:37.481 CC test/nvme/boot_partition/boot_partition.o 00:04:37.481 LINK connect_stress 00:04:37.481 CXX test/cpp_headers/file.o 00:04:37.481 CC examples/blob/hello_world/hello_blob.o 00:04:37.481 CC test/app/histogram_perf/histogram_perf.o 00:04:37.481 CXX test/cpp_headers/fsdev.o 00:04:37.481 LINK boot_partition 00:04:37.481 LINK histogram_perf 00:04:37.481 CC test/nvme/compliance/nvme_compliance.o 00:04:37.481 CC test/app/jsoncat/jsoncat.o 00:04:37.794 LINK hello_world 00:04:37.794 LINK hello_blob 00:04:37.794 CXX test/cpp_headers/fsdev_module.o 00:04:37.794 CXX test/cpp_headers/ftl.o 00:04:37.794 CXX test/cpp_headers/fuse_dispatcher.o 00:04:37.794 LINK jsoncat 00:04:37.794 CXX test/cpp_headers/gpt_spec.o 00:04:37.794 CC examples/nvme/reconnect/reconnect.o 00:04:37.794 LINK spdk_nvme_identify 00:04:37.794 CC test/nvme/fused_ordering/fused_ordering.o 00:04:38.053 CC test/app/stub/stub.o 00:04:38.053 CC examples/blob/cli/blobcli.o 00:04:38.053 LINK nvme_compliance 00:04:38.053 LINK spdk_top 00:04:38.053 CC app/vhost/vhost.o 00:04:38.053 CXX test/cpp_headers/hexlify.o 00:04:38.053 LINK fused_ordering 00:04:38.053 LINK stub 00:04:38.053 CC test/nvme/doorbell_aers/doorbell_aers.o 00:04:38.053 LINK vhost 00:04:38.053 CXX test/cpp_headers/histogram_data.o 00:04:38.312 LINK reconnect 00:04:38.312 CC app/spdk_dd/spdk_dd.o 00:04:38.312 CC app/fio/nvme/fio_plugin.o 00:04:38.312 CXX test/cpp_headers/idxd.o 00:04:38.312 LINK doorbell_aers 00:04:38.312 CC examples/nvme/nvme_manage/nvme_manage.o 00:04:38.572 CC test/nvme/fdp/fdp.o 00:04:38.572 CC app/fio/bdev/fio_plugin.o 00:04:38.572 CC examples/nvme/arbitration/arbitration.o 00:04:38.572 LINK blobcli 00:04:38.572 CXX test/cpp_headers/idxd_spec.o 00:04:38.572 CC test/nvme/cuse/cuse.o 00:04:38.572 LINK spdk_dd 00:04:38.572 CXX test/cpp_headers/init.o 00:04:38.830 CC examples/nvme/hotplug/hotplug.o 00:04:38.830 LINK fdp 00:04:38.830 CXX test/cpp_headers/ioat.o 00:04:38.830 LINK arbitration 00:04:38.830 CC examples/nvme/cmb_copy/cmb_copy.o 00:04:38.830 LINK spdk_nvme 00:04:38.830 LINK nvme_manage 00:04:38.830 LINK spdk_bdev 00:04:39.089 CXX test/cpp_headers/ioat_spec.o 00:04:39.089 LINK hotplug 00:04:39.089 CXX test/cpp_headers/iscsi_spec.o 00:04:39.089 CC examples/nvme/abort/abort.o 00:04:39.089 CXX test/cpp_headers/json.o 00:04:39.089 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:04:39.089 LINK cmb_copy 00:04:39.089 CXX test/cpp_headers/jsonrpc.o 00:04:39.089 CXX test/cpp_headers/keyring.o 00:04:39.089 CXX test/cpp_headers/keyring_module.o 00:04:39.089 CXX test/cpp_headers/likely.o 00:04:39.348 LINK pmr_persistence 00:04:39.348 CXX test/cpp_headers/log.o 00:04:39.348 CC examples/bdev/hello_world/hello_bdev.o 00:04:39.348 CXX test/cpp_headers/lvol.o 00:04:39.348 CXX test/cpp_headers/md5.o 00:04:39.348 CXX test/cpp_headers/memory.o 00:04:39.348 CC examples/bdev/bdevperf/bdevperf.o 00:04:39.348 CXX test/cpp_headers/mmio.o 00:04:39.348 CXX test/cpp_headers/nbd.o 00:04:39.348 CXX test/cpp_headers/net.o 00:04:39.348 LINK abort 00:04:39.348 CXX test/cpp_headers/notify.o 00:04:39.608 CXX test/cpp_headers/nvme.o 00:04:39.608 LINK hello_bdev 00:04:39.608 CXX test/cpp_headers/nvme_intel.o 00:04:39.608 CXX test/cpp_headers/nvme_ocssd.o 00:04:39.608 CXX test/cpp_headers/nvme_ocssd_spec.o 00:04:39.608 CXX test/cpp_headers/nvme_spec.o 00:04:39.608 CXX test/cpp_headers/nvme_zns.o 00:04:39.608 CXX test/cpp_headers/nvmf_cmd.o 00:04:39.608 CXX test/cpp_headers/nvmf_fc_spec.o 00:04:39.608 CXX test/cpp_headers/nvmf.o 00:04:39.608 CXX test/cpp_headers/nvmf_spec.o 00:04:39.608 CXX test/cpp_headers/nvmf_transport.o 00:04:39.608 CXX test/cpp_headers/opal.o 00:04:39.867 CXX test/cpp_headers/opal_spec.o 00:04:39.867 CXX test/cpp_headers/pci_ids.o 00:04:39.867 CXX test/cpp_headers/pipe.o 00:04:39.867 CXX test/cpp_headers/queue.o 00:04:39.867 CXX test/cpp_headers/reduce.o 00:04:39.867 LINK cuse 00:04:39.867 CXX test/cpp_headers/rpc.o 00:04:39.867 CXX test/cpp_headers/scheduler.o 00:04:39.867 CXX test/cpp_headers/scsi.o 00:04:39.867 CXX test/cpp_headers/scsi_spec.o 00:04:39.867 CXX test/cpp_headers/sock.o 00:04:39.867 CXX test/cpp_headers/stdinc.o 00:04:40.125 CXX test/cpp_headers/string.o 00:04:40.125 CXX test/cpp_headers/thread.o 00:04:40.125 CXX test/cpp_headers/trace.o 00:04:40.125 CXX test/cpp_headers/trace_parser.o 00:04:40.125 CXX test/cpp_headers/tree.o 00:04:40.125 CXX test/cpp_headers/ublk.o 00:04:40.125 CXX test/cpp_headers/util.o 00:04:40.125 CXX test/cpp_headers/uuid.o 00:04:40.125 CXX test/cpp_headers/version.o 00:04:40.125 CXX test/cpp_headers/vfio_user_pci.o 00:04:40.125 CXX test/cpp_headers/vfio_user_spec.o 00:04:40.125 CXX test/cpp_headers/vhost.o 00:04:40.125 CXX test/cpp_headers/vmd.o 00:04:40.125 LINK bdevperf 00:04:40.125 CXX test/cpp_headers/xor.o 00:04:40.384 CXX test/cpp_headers/zipf.o 00:04:40.953 CC examples/nvmf/nvmf/nvmf.o 00:04:41.213 LINK nvmf 00:04:41.780 LINK esnap 00:04:42.039 00:04:42.039 real 1m14.894s 00:04:42.039 user 5m55.972s 00:04:42.039 sys 1m7.660s 00:04:42.039 05:32:15 make -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:04:42.039 05:32:15 make -- common/autotest_common.sh@10 -- $ set +x 00:04:42.039 ************************************ 00:04:42.039 END TEST make 00:04:42.039 ************************************ 00:04:42.039 05:32:15 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:04:42.039 05:32:15 -- pm/common@29 -- $ signal_monitor_resources TERM 00:04:42.039 05:32:15 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:04:42.039 05:32:15 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:42.039 05:32:15 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:04:42.039 05:32:15 -- pm/common@44 -- $ pid=6214 00:04:42.039 05:32:15 -- pm/common@50 -- $ kill -TERM 6214 00:04:42.039 05:32:15 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:42.039 05:32:15 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:04:42.039 05:32:15 -- pm/common@44 -- $ pid=6216 00:04:42.039 05:32:15 -- pm/common@50 -- $ kill -TERM 6216 00:04:42.039 05:32:15 -- spdk/autorun.sh@26 -- $ (( SPDK_TEST_UNITTEST == 1 || SPDK_RUN_FUNCTIONAL_TEST == 1 )) 00:04:42.039 05:32:15 -- spdk/autorun.sh@27 -- $ sudo -E /home/vagrant/spdk_repo/spdk/autotest.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:04:42.297 05:32:15 -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:04:42.297 05:32:15 -- common/autotest_common.sh@1711 -- # lcov --version 00:04:42.297 05:32:15 -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:04:42.297 05:32:15 -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:04:42.297 05:32:15 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:42.297 05:32:15 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:42.297 05:32:15 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:42.297 05:32:15 -- scripts/common.sh@336 -- # IFS=.-: 00:04:42.297 05:32:15 -- scripts/common.sh@336 -- # read -ra ver1 00:04:42.297 05:32:15 -- scripts/common.sh@337 -- # IFS=.-: 00:04:42.297 05:32:15 -- scripts/common.sh@337 -- # read -ra ver2 00:04:42.297 05:32:15 -- scripts/common.sh@338 -- # local 'op=<' 00:04:42.297 05:32:15 -- scripts/common.sh@340 -- # ver1_l=2 00:04:42.297 05:32:15 -- scripts/common.sh@341 -- # ver2_l=1 00:04:42.297 05:32:15 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:42.297 05:32:15 -- scripts/common.sh@344 -- # case "$op" in 00:04:42.297 05:32:15 -- scripts/common.sh@345 -- # : 1 00:04:42.297 05:32:15 -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:42.297 05:32:15 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:42.297 05:32:15 -- scripts/common.sh@365 -- # decimal 1 00:04:42.297 05:32:15 -- scripts/common.sh@353 -- # local d=1 00:04:42.297 05:32:15 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:42.297 05:32:15 -- scripts/common.sh@355 -- # echo 1 00:04:42.297 05:32:15 -- scripts/common.sh@365 -- # ver1[v]=1 00:04:42.297 05:32:15 -- scripts/common.sh@366 -- # decimal 2 00:04:42.297 05:32:15 -- scripts/common.sh@353 -- # local d=2 00:04:42.297 05:32:15 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:42.297 05:32:15 -- scripts/common.sh@355 -- # echo 2 00:04:42.297 05:32:15 -- scripts/common.sh@366 -- # ver2[v]=2 00:04:42.297 05:32:15 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:42.297 05:32:15 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:42.297 05:32:15 -- scripts/common.sh@368 -- # return 0 00:04:42.297 05:32:15 -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:42.297 05:32:15 -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:04:42.297 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:42.297 --rc genhtml_branch_coverage=1 00:04:42.297 --rc genhtml_function_coverage=1 00:04:42.297 --rc genhtml_legend=1 00:04:42.297 --rc geninfo_all_blocks=1 00:04:42.297 --rc geninfo_unexecuted_blocks=1 00:04:42.297 00:04:42.297 ' 00:04:42.297 05:32:15 -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:04:42.297 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:42.297 --rc genhtml_branch_coverage=1 00:04:42.297 --rc genhtml_function_coverage=1 00:04:42.297 --rc genhtml_legend=1 00:04:42.297 --rc geninfo_all_blocks=1 00:04:42.297 --rc geninfo_unexecuted_blocks=1 00:04:42.297 00:04:42.297 ' 00:04:42.297 05:32:15 -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:04:42.297 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:42.297 --rc genhtml_branch_coverage=1 00:04:42.297 --rc genhtml_function_coverage=1 00:04:42.297 --rc genhtml_legend=1 00:04:42.297 --rc geninfo_all_blocks=1 00:04:42.297 --rc geninfo_unexecuted_blocks=1 00:04:42.297 00:04:42.297 ' 00:04:42.297 05:32:15 -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:04:42.297 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:42.297 --rc genhtml_branch_coverage=1 00:04:42.297 --rc genhtml_function_coverage=1 00:04:42.297 --rc genhtml_legend=1 00:04:42.297 --rc geninfo_all_blocks=1 00:04:42.297 --rc geninfo_unexecuted_blocks=1 00:04:42.297 00:04:42.297 ' 00:04:42.297 05:32:15 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:04:42.297 05:32:15 -- nvmf/common.sh@7 -- # uname -s 00:04:42.297 05:32:15 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:04:42.297 05:32:15 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:04:42.297 05:32:15 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:04:42.297 05:32:15 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:04:42.297 05:32:15 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:04:42.297 05:32:15 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:04:42.297 05:32:15 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:04:42.297 05:32:15 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:04:42.297 05:32:15 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:04:42.297 05:32:15 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:04:42.297 05:32:15 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:1b6716e3-b0c8-480d-9903-786aaf87e8fc 00:04:42.297 05:32:15 -- nvmf/common.sh@18 -- # NVME_HOSTID=1b6716e3-b0c8-480d-9903-786aaf87e8fc 00:04:42.297 05:32:15 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:04:42.297 05:32:15 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:04:42.297 05:32:15 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:04:42.297 05:32:15 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:04:42.297 05:32:15 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:04:42.297 05:32:15 -- scripts/common.sh@15 -- # shopt -s extglob 00:04:42.297 05:32:15 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:04:42.297 05:32:15 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:42.297 05:32:15 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:42.297 05:32:15 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:42.297 05:32:15 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:42.297 05:32:15 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:42.297 05:32:15 -- paths/export.sh@5 -- # export PATH 00:04:42.297 05:32:15 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:42.297 05:32:15 -- nvmf/common.sh@51 -- # : 0 00:04:42.297 05:32:15 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:04:42.297 05:32:15 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:04:42.297 05:32:15 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:04:42.297 05:32:15 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:04:42.297 05:32:15 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:04:42.297 05:32:15 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:04:42.297 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:04:42.297 05:32:15 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:04:42.297 05:32:15 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:04:42.297 05:32:15 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:04:42.297 05:32:15 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:04:42.297 05:32:15 -- spdk/autotest.sh@32 -- # uname -s 00:04:42.297 05:32:15 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:04:42.297 05:32:15 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:04:42.297 05:32:15 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:04:42.297 05:32:15 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:04:42.297 05:32:15 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:04:42.297 05:32:15 -- spdk/autotest.sh@44 -- # modprobe nbd 00:04:42.297 05:32:15 -- spdk/autotest.sh@46 -- # type -P udevadm 00:04:42.297 05:32:15 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:04:42.297 05:32:15 -- spdk/autotest.sh@48 -- # udevadm_pid=66486 00:04:42.297 05:32:15 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:04:42.297 05:32:15 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:04:42.297 05:32:15 -- pm/common@17 -- # local monitor 00:04:42.297 05:32:15 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:04:42.297 05:32:15 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:04:42.297 05:32:15 -- pm/common@25 -- # sleep 1 00:04:42.297 05:32:15 -- pm/common@21 -- # date +%s 00:04:42.297 05:32:15 -- pm/common@21 -- # date +%s 00:04:42.297 05:32:15 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1733549535 00:04:42.297 05:32:15 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1733549535 00:04:42.297 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1733549535_collect-cpu-load.pm.log 00:04:42.297 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1733549535_collect-vmstat.pm.log 00:04:43.671 05:32:16 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:04:43.671 05:32:16 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:04:43.671 05:32:16 -- common/autotest_common.sh@726 -- # xtrace_disable 00:04:43.671 05:32:16 -- common/autotest_common.sh@10 -- # set +x 00:04:43.671 05:32:16 -- spdk/autotest.sh@59 -- # create_test_list 00:04:43.672 05:32:16 -- common/autotest_common.sh@752 -- # xtrace_disable 00:04:43.672 05:32:16 -- common/autotest_common.sh@10 -- # set +x 00:04:43.672 05:32:16 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:04:43.672 05:32:16 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:04:43.672 05:32:16 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:04:43.672 05:32:16 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:04:43.672 05:32:16 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:04:43.672 05:32:16 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:04:43.672 05:32:16 -- common/autotest_common.sh@1457 -- # uname 00:04:43.672 05:32:16 -- common/autotest_common.sh@1457 -- # '[' Linux = FreeBSD ']' 00:04:43.672 05:32:16 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:04:43.672 05:32:16 -- common/autotest_common.sh@1477 -- # uname 00:04:43.672 05:32:16 -- common/autotest_common.sh@1477 -- # [[ Linux = FreeBSD ]] 00:04:43.672 05:32:16 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:04:43.672 05:32:16 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:04:43.672 lcov: LCOV version 1.15 00:04:43.672 05:32:16 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:04:58.543 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:04:58.543 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:05:13.426 05:32:44 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:05:13.426 05:32:44 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:13.426 05:32:44 -- common/autotest_common.sh@10 -- # set +x 00:05:13.426 05:32:44 -- spdk/autotest.sh@78 -- # rm -f 00:05:13.426 05:32:44 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:13.426 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:13.426 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:05:13.426 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:05:13.426 05:32:45 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:05:13.426 05:32:45 -- common/autotest_common.sh@1657 -- # zoned_devs=() 00:05:13.426 05:32:45 -- common/autotest_common.sh@1657 -- # local -gA zoned_devs 00:05:13.426 05:32:45 -- common/autotest_common.sh@1658 -- # zoned_ctrls=() 00:05:13.427 05:32:45 -- common/autotest_common.sh@1658 -- # local -A zoned_ctrls 00:05:13.427 05:32:45 -- common/autotest_common.sh@1659 -- # local nvme bdf ns 00:05:13.427 05:32:45 -- common/autotest_common.sh@1668 -- # for nvme in /sys/class/nvme/nvme* 00:05:13.427 05:32:45 -- common/autotest_common.sh@1669 -- # bdf=0000:00:10.0 00:05:13.427 05:32:45 -- common/autotest_common.sh@1670 -- # for ns in "$nvme/"nvme*n* 00:05:13.427 05:32:45 -- common/autotest_common.sh@1671 -- # is_block_zoned nvme0n1 00:05:13.427 05:32:45 -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:05:13.427 05:32:45 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:05:13.427 05:32:45 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:13.427 05:32:45 -- common/autotest_common.sh@1668 -- # for nvme in /sys/class/nvme/nvme* 00:05:13.427 05:32:45 -- common/autotest_common.sh@1669 -- # bdf=0000:00:11.0 00:05:13.427 05:32:45 -- common/autotest_common.sh@1670 -- # for ns in "$nvme/"nvme*n* 00:05:13.427 05:32:45 -- common/autotest_common.sh@1671 -- # is_block_zoned nvme1n1 00:05:13.427 05:32:45 -- common/autotest_common.sh@1650 -- # local device=nvme1n1 00:05:13.427 05:32:45 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:05:13.427 05:32:45 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:13.427 05:32:45 -- common/autotest_common.sh@1670 -- # for ns in "$nvme/"nvme*n* 00:05:13.427 05:32:45 -- common/autotest_common.sh@1671 -- # is_block_zoned nvme1n2 00:05:13.427 05:32:45 -- common/autotest_common.sh@1650 -- # local device=nvme1n2 00:05:13.427 05:32:45 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n2/queue/zoned ]] 00:05:13.427 05:32:45 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:13.427 05:32:45 -- common/autotest_common.sh@1670 -- # for ns in "$nvme/"nvme*n* 00:05:13.427 05:32:45 -- common/autotest_common.sh@1671 -- # is_block_zoned nvme1n3 00:05:13.427 05:32:45 -- common/autotest_common.sh@1650 -- # local device=nvme1n3 00:05:13.427 05:32:45 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n3/queue/zoned ]] 00:05:13.427 05:32:45 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:13.427 05:32:45 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:05:13.427 05:32:45 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:13.427 05:32:45 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:13.427 05:32:45 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:05:13.427 05:32:45 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:05:13.427 05:32:45 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:05:13.427 No valid GPT data, bailing 00:05:13.427 05:32:45 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:05:13.427 05:32:45 -- scripts/common.sh@394 -- # pt= 00:05:13.427 05:32:45 -- scripts/common.sh@395 -- # return 1 00:05:13.427 05:32:45 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:05:13.427 1+0 records in 00:05:13.427 1+0 records out 00:05:13.427 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00652048 s, 161 MB/s 00:05:13.427 05:32:45 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:13.427 05:32:45 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:13.427 05:32:45 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:05:13.427 05:32:45 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:05:13.427 05:32:45 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:05:13.427 No valid GPT data, bailing 00:05:13.427 05:32:45 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:05:13.427 05:32:45 -- scripts/common.sh@394 -- # pt= 00:05:13.427 05:32:45 -- scripts/common.sh@395 -- # return 1 00:05:13.427 05:32:45 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:05:13.427 1+0 records in 00:05:13.427 1+0 records out 00:05:13.427 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00404619 s, 259 MB/s 00:05:13.427 05:32:45 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:13.427 05:32:45 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:13.427 05:32:45 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n2 00:05:13.427 05:32:45 -- scripts/common.sh@381 -- # local block=/dev/nvme1n2 pt 00:05:13.427 05:32:45 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n2 00:05:13.427 No valid GPT data, bailing 00:05:13.427 05:32:46 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n2 00:05:13.427 05:32:46 -- scripts/common.sh@394 -- # pt= 00:05:13.427 05:32:46 -- scripts/common.sh@395 -- # return 1 00:05:13.427 05:32:46 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n2 bs=1M count=1 00:05:13.427 1+0 records in 00:05:13.427 1+0 records out 00:05:13.427 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00655525 s, 160 MB/s 00:05:13.427 05:32:46 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:13.427 05:32:46 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:13.427 05:32:46 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n3 00:05:13.427 05:32:46 -- scripts/common.sh@381 -- # local block=/dev/nvme1n3 pt 00:05:13.427 05:32:46 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n3 00:05:13.427 No valid GPT data, bailing 00:05:13.427 05:32:46 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n3 00:05:13.427 05:32:46 -- scripts/common.sh@394 -- # pt= 00:05:13.427 05:32:46 -- scripts/common.sh@395 -- # return 1 00:05:13.427 05:32:46 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n3 bs=1M count=1 00:05:13.427 1+0 records in 00:05:13.427 1+0 records out 00:05:13.427 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00673837 s, 156 MB/s 00:05:13.427 05:32:46 -- spdk/autotest.sh@105 -- # sync 00:05:13.427 05:32:46 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:05:13.427 05:32:46 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:05:13.427 05:32:46 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:05:15.969 05:32:49 -- spdk/autotest.sh@111 -- # uname -s 00:05:15.969 05:32:49 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:05:15.969 05:32:49 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:05:15.969 05:32:49 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:05:16.538 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:16.538 Hugepages 00:05:16.538 node hugesize free / total 00:05:16.538 node0 1048576kB 0 / 0 00:05:16.538 node0 2048kB 0 / 0 00:05:16.538 00:05:16.538 Type BDF Vendor Device NUMA Driver Device Block devices 00:05:16.798 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:05:16.798 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:05:16.798 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:05:16.798 05:32:50 -- spdk/autotest.sh@117 -- # uname -s 00:05:16.798 05:32:50 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:05:16.798 05:32:50 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:05:16.798 05:32:50 -- common/autotest_common.sh@1516 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:17.734 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:17.735 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:17.735 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:17.993 05:32:51 -- common/autotest_common.sh@1517 -- # sleep 1 00:05:18.929 05:32:52 -- common/autotest_common.sh@1518 -- # bdfs=() 00:05:18.929 05:32:52 -- common/autotest_common.sh@1518 -- # local bdfs 00:05:18.929 05:32:52 -- common/autotest_common.sh@1520 -- # bdfs=($(get_nvme_bdfs)) 00:05:18.929 05:32:52 -- common/autotest_common.sh@1520 -- # get_nvme_bdfs 00:05:18.929 05:32:52 -- common/autotest_common.sh@1498 -- # bdfs=() 00:05:18.929 05:32:52 -- common/autotest_common.sh@1498 -- # local bdfs 00:05:18.929 05:32:52 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:18.929 05:32:52 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:18.929 05:32:52 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:05:18.929 05:32:52 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:05:18.929 05:32:52 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:18.929 05:32:52 -- common/autotest_common.sh@1522 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:19.497 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:19.497 Waiting for block devices as requested 00:05:19.497 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:05:19.497 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:05:19.755 05:32:52 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:05:19.756 05:32:52 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:05:19.756 05:32:52 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:19.756 05:32:52 -- common/autotest_common.sh@1487 -- # grep 0000:00:10.0/nvme/nvme 00:05:19.756 05:32:52 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:19.756 05:32:52 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:05:19.756 05:32:52 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:19.756 05:32:52 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme1 00:05:19.756 05:32:52 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme1 00:05:19.756 05:32:52 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme1 ]] 00:05:19.756 05:32:52 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme1 00:05:19.756 05:32:52 -- common/autotest_common.sh@1531 -- # grep oacs 00:05:19.756 05:32:52 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:05:19.756 05:32:52 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:05:19.756 05:32:52 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:05:19.756 05:32:52 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:05:19.756 05:32:52 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme1 00:05:19.756 05:32:52 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:05:19.756 05:32:52 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:05:19.756 05:32:52 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:05:19.756 05:32:52 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:05:19.756 05:32:52 -- common/autotest_common.sh@1543 -- # continue 00:05:19.756 05:32:52 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:05:19.756 05:32:52 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:05:19.756 05:32:52 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:19.756 05:32:52 -- common/autotest_common.sh@1487 -- # grep 0000:00:11.0/nvme/nvme 00:05:19.756 05:32:53 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:19.756 05:32:53 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:05:19.756 05:32:53 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:19.756 05:32:53 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme0 00:05:19.756 05:32:53 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme0 00:05:19.756 05:32:53 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme0 ]] 00:05:19.756 05:32:53 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme0 00:05:19.756 05:32:53 -- common/autotest_common.sh@1531 -- # grep oacs 00:05:19.756 05:32:53 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:05:19.756 05:32:53 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:05:19.756 05:32:53 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:05:19.756 05:32:53 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:05:19.756 05:32:53 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme0 00:05:19.756 05:32:53 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:05:19.756 05:32:53 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:05:19.756 05:32:53 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:05:19.756 05:32:53 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:05:19.756 05:32:53 -- common/autotest_common.sh@1543 -- # continue 00:05:19.756 05:32:53 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:05:19.756 05:32:53 -- common/autotest_common.sh@732 -- # xtrace_disable 00:05:19.756 05:32:53 -- common/autotest_common.sh@10 -- # set +x 00:05:19.756 05:32:53 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:05:19.756 05:32:53 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:19.756 05:32:53 -- common/autotest_common.sh@10 -- # set +x 00:05:19.756 05:32:53 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:20.692 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:20.692 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:20.692 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:20.951 05:32:54 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:05:20.951 05:32:54 -- common/autotest_common.sh@732 -- # xtrace_disable 00:05:20.951 05:32:54 -- common/autotest_common.sh@10 -- # set +x 00:05:20.951 05:32:54 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:05:20.951 05:32:54 -- common/autotest_common.sh@1578 -- # mapfile -t bdfs 00:05:20.951 05:32:54 -- common/autotest_common.sh@1578 -- # get_nvme_bdfs_by_id 0x0a54 00:05:20.951 05:32:54 -- common/autotest_common.sh@1563 -- # bdfs=() 00:05:20.951 05:32:54 -- common/autotest_common.sh@1563 -- # _bdfs=() 00:05:20.951 05:32:54 -- common/autotest_common.sh@1563 -- # local bdfs _bdfs 00:05:20.951 05:32:54 -- common/autotest_common.sh@1564 -- # _bdfs=($(get_nvme_bdfs)) 00:05:20.951 05:32:54 -- common/autotest_common.sh@1564 -- # get_nvme_bdfs 00:05:20.951 05:32:54 -- common/autotest_common.sh@1498 -- # bdfs=() 00:05:20.951 05:32:54 -- common/autotest_common.sh@1498 -- # local bdfs 00:05:20.951 05:32:54 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:20.951 05:32:54 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:20.951 05:32:54 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:05:20.951 05:32:54 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:05:20.951 05:32:54 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:20.951 05:32:54 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:05:20.951 05:32:54 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:05:20.951 05:32:54 -- common/autotest_common.sh@1566 -- # device=0x0010 00:05:20.951 05:32:54 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:20.951 05:32:54 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:05:20.951 05:32:54 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:05:20.951 05:32:54 -- common/autotest_common.sh@1566 -- # device=0x0010 00:05:20.951 05:32:54 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:20.951 05:32:54 -- common/autotest_common.sh@1572 -- # (( 0 > 0 )) 00:05:20.951 05:32:54 -- common/autotest_common.sh@1572 -- # return 0 00:05:20.951 05:32:54 -- common/autotest_common.sh@1579 -- # [[ -z '' ]] 00:05:20.951 05:32:54 -- common/autotest_common.sh@1580 -- # return 0 00:05:20.951 05:32:54 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:05:20.951 05:32:54 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:05:20.951 05:32:54 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:20.951 05:32:54 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:20.951 05:32:54 -- spdk/autotest.sh@149 -- # timing_enter lib 00:05:20.951 05:32:54 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:20.951 05:32:54 -- common/autotest_common.sh@10 -- # set +x 00:05:20.951 05:32:54 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:05:20.951 05:32:54 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:20.951 05:32:54 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:20.951 05:32:54 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:20.951 05:32:54 -- common/autotest_common.sh@10 -- # set +x 00:05:20.951 ************************************ 00:05:20.951 START TEST env 00:05:20.951 ************************************ 00:05:20.951 05:32:54 env -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:21.210 * Looking for test storage... 00:05:21.210 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:05:21.210 05:32:54 env -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:21.210 05:32:54 env -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:21.210 05:32:54 env -- common/autotest_common.sh@1711 -- # lcov --version 00:05:21.210 05:32:54 env -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:21.210 05:32:54 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:21.210 05:32:54 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:21.210 05:32:54 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:21.210 05:32:54 env -- scripts/common.sh@336 -- # IFS=.-: 00:05:21.210 05:32:54 env -- scripts/common.sh@336 -- # read -ra ver1 00:05:21.210 05:32:54 env -- scripts/common.sh@337 -- # IFS=.-: 00:05:21.210 05:32:54 env -- scripts/common.sh@337 -- # read -ra ver2 00:05:21.210 05:32:54 env -- scripts/common.sh@338 -- # local 'op=<' 00:05:21.210 05:32:54 env -- scripts/common.sh@340 -- # ver1_l=2 00:05:21.210 05:32:54 env -- scripts/common.sh@341 -- # ver2_l=1 00:05:21.210 05:32:54 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:21.210 05:32:54 env -- scripts/common.sh@344 -- # case "$op" in 00:05:21.210 05:32:54 env -- scripts/common.sh@345 -- # : 1 00:05:21.210 05:32:54 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:21.210 05:32:54 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:21.210 05:32:54 env -- scripts/common.sh@365 -- # decimal 1 00:05:21.210 05:32:54 env -- scripts/common.sh@353 -- # local d=1 00:05:21.210 05:32:54 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:21.210 05:32:54 env -- scripts/common.sh@355 -- # echo 1 00:05:21.210 05:32:54 env -- scripts/common.sh@365 -- # ver1[v]=1 00:05:21.210 05:32:54 env -- scripts/common.sh@366 -- # decimal 2 00:05:21.210 05:32:54 env -- scripts/common.sh@353 -- # local d=2 00:05:21.210 05:32:54 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:21.210 05:32:54 env -- scripts/common.sh@355 -- # echo 2 00:05:21.210 05:32:54 env -- scripts/common.sh@366 -- # ver2[v]=2 00:05:21.210 05:32:54 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:21.210 05:32:54 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:21.210 05:32:54 env -- scripts/common.sh@368 -- # return 0 00:05:21.210 05:32:54 env -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:21.210 05:32:54 env -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:21.210 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:21.210 --rc genhtml_branch_coverage=1 00:05:21.210 --rc genhtml_function_coverage=1 00:05:21.210 --rc genhtml_legend=1 00:05:21.210 --rc geninfo_all_blocks=1 00:05:21.210 --rc geninfo_unexecuted_blocks=1 00:05:21.210 00:05:21.210 ' 00:05:21.210 05:32:54 env -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:21.210 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:21.210 --rc genhtml_branch_coverage=1 00:05:21.210 --rc genhtml_function_coverage=1 00:05:21.210 --rc genhtml_legend=1 00:05:21.210 --rc geninfo_all_blocks=1 00:05:21.210 --rc geninfo_unexecuted_blocks=1 00:05:21.210 00:05:21.210 ' 00:05:21.210 05:32:54 env -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:21.210 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:21.210 --rc genhtml_branch_coverage=1 00:05:21.210 --rc genhtml_function_coverage=1 00:05:21.210 --rc genhtml_legend=1 00:05:21.210 --rc geninfo_all_blocks=1 00:05:21.210 --rc geninfo_unexecuted_blocks=1 00:05:21.210 00:05:21.210 ' 00:05:21.210 05:32:54 env -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:21.210 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:21.210 --rc genhtml_branch_coverage=1 00:05:21.210 --rc genhtml_function_coverage=1 00:05:21.210 --rc genhtml_legend=1 00:05:21.210 --rc geninfo_all_blocks=1 00:05:21.210 --rc geninfo_unexecuted_blocks=1 00:05:21.210 00:05:21.210 ' 00:05:21.210 05:32:54 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:21.210 05:32:54 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:21.210 05:32:54 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:21.210 05:32:54 env -- common/autotest_common.sh@10 -- # set +x 00:05:21.210 ************************************ 00:05:21.210 START TEST env_memory 00:05:21.210 ************************************ 00:05:21.210 05:32:54 env.env_memory -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:21.210 00:05:21.210 00:05:21.210 CUnit - A unit testing framework for C - Version 2.1-3 00:05:21.210 http://cunit.sourceforge.net/ 00:05:21.210 00:05:21.210 00:05:21.210 Suite: memory 00:05:21.469 Test: alloc and free memory map ...[2024-12-07 05:32:54.586487] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:05:21.469 passed 00:05:21.469 Test: mem map translation ...[2024-12-07 05:32:54.629162] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:05:21.469 [2024-12-07 05:32:54.629241] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:05:21.469 [2024-12-07 05:32:54.629304] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:05:21.469 [2024-12-07 05:32:54.629324] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:05:21.469 passed 00:05:21.469 Test: mem map registration ...[2024-12-07 05:32:54.694579] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:05:21.469 [2024-12-07 05:32:54.694650] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:05:21.469 passed 00:05:21.469 Test: mem map adjacent registrations ...passed 00:05:21.469 00:05:21.469 Run Summary: Type Total Ran Passed Failed Inactive 00:05:21.469 suites 1 1 n/a 0 0 00:05:21.469 tests 4 4 4 0 0 00:05:21.469 asserts 152 152 152 0 n/a 00:05:21.469 00:05:21.469 Elapsed time = 0.239 seconds 00:05:21.469 00:05:21.469 real 0m0.297s 00:05:21.469 user 0m0.250s 00:05:21.469 sys 0m0.037s 00:05:21.469 ************************************ 00:05:21.469 END TEST env_memory 00:05:21.469 ************************************ 00:05:21.469 05:32:54 env.env_memory -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:21.469 05:32:54 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:05:21.728 05:32:54 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:21.729 05:32:54 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:21.729 05:32:54 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:21.729 05:32:54 env -- common/autotest_common.sh@10 -- # set +x 00:05:21.729 ************************************ 00:05:21.729 START TEST env_vtophys 00:05:21.729 ************************************ 00:05:21.729 05:32:54 env.env_vtophys -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:21.729 EAL: lib.eal log level changed from notice to debug 00:05:21.729 EAL: Detected lcore 0 as core 0 on socket 0 00:05:21.729 EAL: Detected lcore 1 as core 0 on socket 0 00:05:21.729 EAL: Detected lcore 2 as core 0 on socket 0 00:05:21.729 EAL: Detected lcore 3 as core 0 on socket 0 00:05:21.729 EAL: Detected lcore 4 as core 0 on socket 0 00:05:21.729 EAL: Detected lcore 5 as core 0 on socket 0 00:05:21.729 EAL: Detected lcore 6 as core 0 on socket 0 00:05:21.729 EAL: Detected lcore 7 as core 0 on socket 0 00:05:21.729 EAL: Detected lcore 8 as core 0 on socket 0 00:05:21.729 EAL: Detected lcore 9 as core 0 on socket 0 00:05:21.729 EAL: Maximum logical cores by configuration: 128 00:05:21.729 EAL: Detected CPU lcores: 10 00:05:21.729 EAL: Detected NUMA nodes: 1 00:05:21.729 EAL: Checking presence of .so 'librte_eal.so.23.0' 00:05:21.729 EAL: Detected shared linkage of DPDK 00:05:21.729 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23.0 00:05:21.729 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23.0 00:05:21.729 EAL: Registered [vdev] bus. 00:05:21.729 EAL: bus.vdev log level changed from disabled to notice 00:05:21.729 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23.0 00:05:21.729 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23.0 00:05:21.729 EAL: pmd.net.i40e.init log level changed from disabled to notice 00:05:21.729 EAL: pmd.net.i40e.driver log level changed from disabled to notice 00:05:21.729 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:05:21.729 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:05:21.729 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:05:21.729 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:05:21.729 EAL: No shared files mode enabled, IPC will be disabled 00:05:21.729 EAL: No shared files mode enabled, IPC is disabled 00:05:21.729 EAL: Selected IOVA mode 'PA' 00:05:21.729 EAL: Probing VFIO support... 00:05:21.729 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:21.729 EAL: VFIO modules not loaded, skipping VFIO support... 00:05:21.729 EAL: Ask a virtual area of 0x2e000 bytes 00:05:21.729 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:05:21.729 EAL: Setting up physically contiguous memory... 00:05:21.729 EAL: Setting maximum number of open files to 524288 00:05:21.729 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:05:21.729 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:05:21.729 EAL: Ask a virtual area of 0x61000 bytes 00:05:21.729 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:05:21.729 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:21.729 EAL: Ask a virtual area of 0x400000000 bytes 00:05:21.729 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:05:21.729 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:05:21.729 EAL: Ask a virtual area of 0x61000 bytes 00:05:21.729 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:05:21.729 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:21.729 EAL: Ask a virtual area of 0x400000000 bytes 00:05:21.729 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:05:21.729 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:05:21.729 EAL: Ask a virtual area of 0x61000 bytes 00:05:21.729 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:05:21.729 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:21.729 EAL: Ask a virtual area of 0x400000000 bytes 00:05:21.729 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:05:21.729 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:05:21.729 EAL: Ask a virtual area of 0x61000 bytes 00:05:21.729 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:05:21.729 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:21.729 EAL: Ask a virtual area of 0x400000000 bytes 00:05:21.729 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:05:21.729 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:05:21.729 EAL: Hugepages will be freed exactly as allocated. 00:05:21.729 EAL: No shared files mode enabled, IPC is disabled 00:05:21.729 EAL: No shared files mode enabled, IPC is disabled 00:05:21.729 EAL: TSC frequency is ~2290000 KHz 00:05:21.729 EAL: Main lcore 0 is ready (tid=7f99f5815a40;cpuset=[0]) 00:05:21.729 EAL: Trying to obtain current memory policy. 00:05:21.729 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:21.729 EAL: Restoring previous memory policy: 0 00:05:21.729 EAL: request: mp_malloc_sync 00:05:21.729 EAL: No shared files mode enabled, IPC is disabled 00:05:21.729 EAL: Heap on socket 0 was expanded by 2MB 00:05:21.729 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:21.729 EAL: No shared files mode enabled, IPC is disabled 00:05:21.729 EAL: No PCI address specified using 'addr=' in: bus=pci 00:05:21.729 EAL: Mem event callback 'spdk:(nil)' registered 00:05:21.729 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:05:21.729 00:05:21.729 00:05:21.729 CUnit - A unit testing framework for C - Version 2.1-3 00:05:21.729 http://cunit.sourceforge.net/ 00:05:21.729 00:05:21.729 00:05:21.729 Suite: components_suite 00:05:22.295 Test: vtophys_malloc_test ...passed 00:05:22.295 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:05:22.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.295 EAL: Restoring previous memory policy: 4 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was expanded by 4MB 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was shrunk by 4MB 00:05:22.295 EAL: Trying to obtain current memory policy. 00:05:22.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.295 EAL: Restoring previous memory policy: 4 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was expanded by 6MB 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was shrunk by 6MB 00:05:22.295 EAL: Trying to obtain current memory policy. 00:05:22.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.295 EAL: Restoring previous memory policy: 4 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was expanded by 10MB 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was shrunk by 10MB 00:05:22.295 EAL: Trying to obtain current memory policy. 00:05:22.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.295 EAL: Restoring previous memory policy: 4 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was expanded by 18MB 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was shrunk by 18MB 00:05:22.295 EAL: Trying to obtain current memory policy. 00:05:22.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.295 EAL: Restoring previous memory policy: 4 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was expanded by 34MB 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was shrunk by 34MB 00:05:22.295 EAL: Trying to obtain current memory policy. 00:05:22.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.295 EAL: Restoring previous memory policy: 4 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was expanded by 66MB 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was shrunk by 66MB 00:05:22.295 EAL: Trying to obtain current memory policy. 00:05:22.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.295 EAL: Restoring previous memory policy: 4 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was expanded by 130MB 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was shrunk by 130MB 00:05:22.295 EAL: Trying to obtain current memory policy. 00:05:22.295 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.295 EAL: Restoring previous memory policy: 4 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.295 EAL: request: mp_malloc_sync 00:05:22.295 EAL: No shared files mode enabled, IPC is disabled 00:05:22.295 EAL: Heap on socket 0 was expanded by 258MB 00:05:22.295 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.556 EAL: request: mp_malloc_sync 00:05:22.556 EAL: No shared files mode enabled, IPC is disabled 00:05:22.556 EAL: Heap on socket 0 was shrunk by 258MB 00:05:22.556 EAL: Trying to obtain current memory policy. 00:05:22.556 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.556 EAL: Restoring previous memory policy: 4 00:05:22.556 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.556 EAL: request: mp_malloc_sync 00:05:22.556 EAL: No shared files mode enabled, IPC is disabled 00:05:22.556 EAL: Heap on socket 0 was expanded by 514MB 00:05:22.556 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.820 EAL: request: mp_malloc_sync 00:05:22.820 EAL: No shared files mode enabled, IPC is disabled 00:05:22.820 EAL: Heap on socket 0 was shrunk by 514MB 00:05:22.820 EAL: Trying to obtain current memory policy. 00:05:22.820 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.820 EAL: Restoring previous memory policy: 4 00:05:22.820 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.820 EAL: request: mp_malloc_sync 00:05:22.820 EAL: No shared files mode enabled, IPC is disabled 00:05:22.820 EAL: Heap on socket 0 was expanded by 1026MB 00:05:23.079 EAL: Calling mem event callback 'spdk:(nil)' 00:05:23.338 passed 00:05:23.338 00:05:23.338 Run Summary: Type Total Ran Passed Failed Inactive 00:05:23.338 suites 1 1 n/a 0 0 00:05:23.338 tests 2 2 2 0 0 00:05:23.338 asserts 5232 5232 5232 0 n/a 00:05:23.338 00:05:23.338 Elapsed time = 1.382 seconds 00:05:23.338 EAL: request: mp_malloc_sync 00:05:23.338 EAL: No shared files mode enabled, IPC is disabled 00:05:23.338 EAL: Heap on socket 0 was shrunk by 1026MB 00:05:23.338 EAL: Calling mem event callback 'spdk:(nil)' 00:05:23.338 EAL: request: mp_malloc_sync 00:05:23.338 EAL: No shared files mode enabled, IPC is disabled 00:05:23.338 EAL: Heap on socket 0 was shrunk by 2MB 00:05:23.338 EAL: No shared files mode enabled, IPC is disabled 00:05:23.338 EAL: No shared files mode enabled, IPC is disabled 00:05:23.338 EAL: No shared files mode enabled, IPC is disabled 00:05:23.338 00:05:23.338 real 0m1.641s 00:05:23.338 user 0m0.773s 00:05:23.338 sys 0m0.731s 00:05:23.338 ************************************ 00:05:23.338 END TEST env_vtophys 00:05:23.338 ************************************ 00:05:23.338 05:32:56 env.env_vtophys -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:23.338 05:32:56 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:05:23.338 05:32:56 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:23.338 05:32:56 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:23.338 05:32:56 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:23.338 05:32:56 env -- common/autotest_common.sh@10 -- # set +x 00:05:23.338 ************************************ 00:05:23.338 START TEST env_pci 00:05:23.338 ************************************ 00:05:23.338 05:32:56 env.env_pci -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:23.338 00:05:23.338 00:05:23.338 CUnit - A unit testing framework for C - Version 2.1-3 00:05:23.338 http://cunit.sourceforge.net/ 00:05:23.338 00:05:23.338 00:05:23.338 Suite: pci 00:05:23.338 Test: pci_hook ...[2024-12-07 05:32:56.596600] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1117:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 68708 has claimed it 00:05:23.338 passed 00:05:23.338 00:05:23.338 Run Summary: Type Total Ran Passed Failed Inactive 00:05:23.338 suites 1 1 n/a 0 0 00:05:23.338 tests 1 1 1 0 0 00:05:23.338 asserts 25 25 25 0 n/a 00:05:23.338 00:05:23.338 Elapsed time = 0.007 seconds 00:05:23.338 EAL: Cannot find device (10000:00:01.0) 00:05:23.338 EAL: Failed to attach device on primary process 00:05:23.338 ************************************ 00:05:23.338 END TEST env_pci 00:05:23.338 ************************************ 00:05:23.338 00:05:23.338 real 0m0.090s 00:05:23.338 user 0m0.035s 00:05:23.338 sys 0m0.055s 00:05:23.338 05:32:56 env.env_pci -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:23.338 05:32:56 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:05:23.597 05:32:56 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:05:23.597 05:32:56 env -- env/env.sh@15 -- # uname 00:05:23.597 05:32:56 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:05:23.597 05:32:56 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:05:23.597 05:32:56 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:23.597 05:32:56 env -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:05:23.597 05:32:56 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:23.597 05:32:56 env -- common/autotest_common.sh@10 -- # set +x 00:05:23.597 ************************************ 00:05:23.597 START TEST env_dpdk_post_init 00:05:23.597 ************************************ 00:05:23.597 05:32:56 env.env_dpdk_post_init -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:23.597 EAL: Detected CPU lcores: 10 00:05:23.597 EAL: Detected NUMA nodes: 1 00:05:23.597 EAL: Detected shared linkage of DPDK 00:05:23.597 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:23.597 EAL: Selected IOVA mode 'PA' 00:05:23.597 TELEMETRY: No legacy callbacks, legacy socket not created 00:05:23.597 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:10.0 (socket -1) 00:05:23.597 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:11.0 (socket -1) 00:05:23.597 Starting DPDK initialization... 00:05:23.597 Starting SPDK post initialization... 00:05:23.597 SPDK NVMe probe 00:05:23.597 Attaching to 0000:00:10.0 00:05:23.597 Attaching to 0000:00:11.0 00:05:23.597 Attached to 0000:00:10.0 00:05:23.597 Attached to 0000:00:11.0 00:05:23.597 Cleaning up... 00:05:23.856 00:05:23.856 real 0m0.244s 00:05:23.856 user 0m0.071s 00:05:23.856 sys 0m0.073s 00:05:23.856 05:32:56 env.env_dpdk_post_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:23.856 05:32:56 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:05:23.856 ************************************ 00:05:23.856 END TEST env_dpdk_post_init 00:05:23.856 ************************************ 00:05:23.856 05:32:57 env -- env/env.sh@26 -- # uname 00:05:23.856 05:32:57 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:05:23.856 05:32:57 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:23.856 05:32:57 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:23.856 05:32:57 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:23.856 05:32:57 env -- common/autotest_common.sh@10 -- # set +x 00:05:23.856 ************************************ 00:05:23.856 START TEST env_mem_callbacks 00:05:23.856 ************************************ 00:05:23.856 05:32:57 env.env_mem_callbacks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:23.856 EAL: Detected CPU lcores: 10 00:05:23.856 EAL: Detected NUMA nodes: 1 00:05:23.856 EAL: Detected shared linkage of DPDK 00:05:23.856 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:23.856 EAL: Selected IOVA mode 'PA' 00:05:23.856 TELEMETRY: No legacy callbacks, legacy socket not created 00:05:23.856 00:05:23.856 00:05:23.856 CUnit - A unit testing framework for C - Version 2.1-3 00:05:23.856 http://cunit.sourceforge.net/ 00:05:23.856 00:05:23.856 00:05:23.856 Suite: memory 00:05:23.856 Test: test ... 00:05:23.856 register 0x200000200000 2097152 00:05:23.856 malloc 3145728 00:05:23.856 register 0x200000400000 4194304 00:05:23.856 buf 0x200000500000 len 3145728 PASSED 00:05:23.856 malloc 64 00:05:23.856 buf 0x2000004fff40 len 64 PASSED 00:05:23.856 malloc 4194304 00:05:23.856 register 0x200000800000 6291456 00:05:23.856 buf 0x200000a00000 len 4194304 PASSED 00:05:23.856 free 0x200000500000 3145728 00:05:23.856 free 0x2000004fff40 64 00:05:23.856 unregister 0x200000400000 4194304 PASSED 00:05:23.856 free 0x200000a00000 4194304 00:05:23.856 unregister 0x200000800000 6291456 PASSED 00:05:23.856 malloc 8388608 00:05:23.856 register 0x200000400000 10485760 00:05:23.856 buf 0x200000600000 len 8388608 PASSED 00:05:23.856 free 0x200000600000 8388608 00:05:23.856 unregister 0x200000400000 10485760 PASSED 00:05:23.856 passed 00:05:23.856 00:05:23.856 Run Summary: Type Total Ran Passed Failed Inactive 00:05:23.856 suites 1 1 n/a 0 0 00:05:23.856 tests 1 1 1 0 0 00:05:23.856 asserts 15 15 15 0 n/a 00:05:23.856 00:05:23.856 Elapsed time = 0.012 seconds 00:05:24.115 00:05:24.115 real 0m0.185s 00:05:24.115 user 0m0.031s 00:05:24.115 sys 0m0.053s 00:05:24.115 05:32:57 env.env_mem_callbacks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:24.115 05:32:57 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:05:24.115 ************************************ 00:05:24.115 END TEST env_mem_callbacks 00:05:24.115 ************************************ 00:05:24.115 00:05:24.115 real 0m3.001s 00:05:24.115 user 0m1.372s 00:05:24.115 sys 0m1.296s 00:05:24.115 05:32:57 env -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:24.115 05:32:57 env -- common/autotest_common.sh@10 -- # set +x 00:05:24.115 ************************************ 00:05:24.115 END TEST env 00:05:24.115 ************************************ 00:05:24.115 05:32:57 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:24.115 05:32:57 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:24.115 05:32:57 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:24.115 05:32:57 -- common/autotest_common.sh@10 -- # set +x 00:05:24.116 ************************************ 00:05:24.116 START TEST rpc 00:05:24.116 ************************************ 00:05:24.116 05:32:57 rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:24.116 * Looking for test storage... 00:05:24.116 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:24.116 05:32:57 rpc -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:24.116 05:32:57 rpc -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:24.116 05:32:57 rpc -- common/autotest_common.sh@1711 -- # lcov --version 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:24.375 05:32:57 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:24.375 05:32:57 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:24.375 05:32:57 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:24.375 05:32:57 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:24.375 05:32:57 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:24.375 05:32:57 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:24.375 05:32:57 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:24.375 05:32:57 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:24.375 05:32:57 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:24.375 05:32:57 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:24.375 05:32:57 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:24.375 05:32:57 rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:24.375 05:32:57 rpc -- scripts/common.sh@345 -- # : 1 00:05:24.375 05:32:57 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:24.375 05:32:57 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:24.375 05:32:57 rpc -- scripts/common.sh@365 -- # decimal 1 00:05:24.375 05:32:57 rpc -- scripts/common.sh@353 -- # local d=1 00:05:24.375 05:32:57 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:24.375 05:32:57 rpc -- scripts/common.sh@355 -- # echo 1 00:05:24.375 05:32:57 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:24.375 05:32:57 rpc -- scripts/common.sh@366 -- # decimal 2 00:05:24.375 05:32:57 rpc -- scripts/common.sh@353 -- # local d=2 00:05:24.375 05:32:57 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:24.375 05:32:57 rpc -- scripts/common.sh@355 -- # echo 2 00:05:24.375 05:32:57 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:24.375 05:32:57 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:24.375 05:32:57 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:24.375 05:32:57 rpc -- scripts/common.sh@368 -- # return 0 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:24.375 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:24.375 --rc genhtml_branch_coverage=1 00:05:24.375 --rc genhtml_function_coverage=1 00:05:24.375 --rc genhtml_legend=1 00:05:24.375 --rc geninfo_all_blocks=1 00:05:24.375 --rc geninfo_unexecuted_blocks=1 00:05:24.375 00:05:24.375 ' 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:24.375 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:24.375 --rc genhtml_branch_coverage=1 00:05:24.375 --rc genhtml_function_coverage=1 00:05:24.375 --rc genhtml_legend=1 00:05:24.375 --rc geninfo_all_blocks=1 00:05:24.375 --rc geninfo_unexecuted_blocks=1 00:05:24.375 00:05:24.375 ' 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:24.375 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:24.375 --rc genhtml_branch_coverage=1 00:05:24.375 --rc genhtml_function_coverage=1 00:05:24.375 --rc genhtml_legend=1 00:05:24.375 --rc geninfo_all_blocks=1 00:05:24.375 --rc geninfo_unexecuted_blocks=1 00:05:24.375 00:05:24.375 ' 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:24.375 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:24.375 --rc genhtml_branch_coverage=1 00:05:24.375 --rc genhtml_function_coverage=1 00:05:24.375 --rc genhtml_legend=1 00:05:24.375 --rc geninfo_all_blocks=1 00:05:24.375 --rc geninfo_unexecuted_blocks=1 00:05:24.375 00:05:24.375 ' 00:05:24.375 05:32:57 rpc -- rpc/rpc.sh@65 -- # spdk_pid=68835 00:05:24.375 05:32:57 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:05:24.375 05:32:57 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:24.375 05:32:57 rpc -- rpc/rpc.sh@67 -- # waitforlisten 68835 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@835 -- # '[' -z 68835 ']' 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:24.375 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:24.375 05:32:57 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:24.375 [2024-12-07 05:32:57.651806] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:24.375 [2024-12-07 05:32:57.651975] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid68835 ] 00:05:24.634 [2024-12-07 05:32:57.810140] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:24.634 [2024-12-07 05:32:57.838764] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:05:24.634 [2024-12-07 05:32:57.838830] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 68835' to capture a snapshot of events at runtime. 00:05:24.634 [2024-12-07 05:32:57.838842] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:05:24.634 [2024-12-07 05:32:57.838859] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:05:24.634 [2024-12-07 05:32:57.838871] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid68835 for offline analysis/debug. 00:05:24.634 [2024-12-07 05:32:57.839300] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:25.200 05:32:58 rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:25.200 05:32:58 rpc -- common/autotest_common.sh@868 -- # return 0 00:05:25.200 05:32:58 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:25.200 05:32:58 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:25.200 05:32:58 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:05:25.200 05:32:58 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:05:25.200 05:32:58 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:25.200 05:32:58 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:25.200 05:32:58 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:25.200 ************************************ 00:05:25.200 START TEST rpc_integrity 00:05:25.200 ************************************ 00:05:25.200 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:05:25.200 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:25.200 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.200 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.200 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.200 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:25.200 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:25.460 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:25.460 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.460 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:05:25.460 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.460 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:25.460 { 00:05:25.460 "name": "Malloc0", 00:05:25.460 "aliases": [ 00:05:25.460 "1d10ca61-7c53-4e67-9e06-ba96aa230d37" 00:05:25.460 ], 00:05:25.460 "product_name": "Malloc disk", 00:05:25.460 "block_size": 512, 00:05:25.460 "num_blocks": 16384, 00:05:25.460 "uuid": "1d10ca61-7c53-4e67-9e06-ba96aa230d37", 00:05:25.460 "assigned_rate_limits": { 00:05:25.460 "rw_ios_per_sec": 0, 00:05:25.460 "rw_mbytes_per_sec": 0, 00:05:25.460 "r_mbytes_per_sec": 0, 00:05:25.460 "w_mbytes_per_sec": 0 00:05:25.460 }, 00:05:25.460 "claimed": false, 00:05:25.460 "zoned": false, 00:05:25.460 "supported_io_types": { 00:05:25.460 "read": true, 00:05:25.460 "write": true, 00:05:25.460 "unmap": true, 00:05:25.460 "flush": true, 00:05:25.460 "reset": true, 00:05:25.460 "nvme_admin": false, 00:05:25.460 "nvme_io": false, 00:05:25.460 "nvme_io_md": false, 00:05:25.460 "write_zeroes": true, 00:05:25.460 "zcopy": true, 00:05:25.460 "get_zone_info": false, 00:05:25.460 "zone_management": false, 00:05:25.460 "zone_append": false, 00:05:25.460 "compare": false, 00:05:25.460 "compare_and_write": false, 00:05:25.460 "abort": true, 00:05:25.460 "seek_hole": false, 00:05:25.460 "seek_data": false, 00:05:25.460 "copy": true, 00:05:25.460 "nvme_iov_md": false 00:05:25.460 }, 00:05:25.460 "memory_domains": [ 00:05:25.460 { 00:05:25.460 "dma_device_id": "system", 00:05:25.460 "dma_device_type": 1 00:05:25.460 }, 00:05:25.460 { 00:05:25.460 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:25.460 "dma_device_type": 2 00:05:25.460 } 00:05:25.460 ], 00:05:25.460 "driver_specific": {} 00:05:25.460 } 00:05:25.460 ]' 00:05:25.460 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:25.460 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:25.460 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.460 [2024-12-07 05:32:58.666086] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:05:25.460 [2024-12-07 05:32:58.666187] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:25.460 [2024-12-07 05:32:58.666231] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006c80 00:05:25.460 [2024-12-07 05:32:58.666245] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:25.460 [2024-12-07 05:32:58.669122] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:25.460 [2024-12-07 05:32:58.669171] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:25.460 Passthru0 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.460 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.460 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.460 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:25.460 { 00:05:25.460 "name": "Malloc0", 00:05:25.460 "aliases": [ 00:05:25.460 "1d10ca61-7c53-4e67-9e06-ba96aa230d37" 00:05:25.460 ], 00:05:25.460 "product_name": "Malloc disk", 00:05:25.460 "block_size": 512, 00:05:25.460 "num_blocks": 16384, 00:05:25.460 "uuid": "1d10ca61-7c53-4e67-9e06-ba96aa230d37", 00:05:25.460 "assigned_rate_limits": { 00:05:25.460 "rw_ios_per_sec": 0, 00:05:25.460 "rw_mbytes_per_sec": 0, 00:05:25.460 "r_mbytes_per_sec": 0, 00:05:25.460 "w_mbytes_per_sec": 0 00:05:25.460 }, 00:05:25.460 "claimed": true, 00:05:25.460 "claim_type": "exclusive_write", 00:05:25.460 "zoned": false, 00:05:25.460 "supported_io_types": { 00:05:25.460 "read": true, 00:05:25.460 "write": true, 00:05:25.460 "unmap": true, 00:05:25.460 "flush": true, 00:05:25.460 "reset": true, 00:05:25.460 "nvme_admin": false, 00:05:25.460 "nvme_io": false, 00:05:25.460 "nvme_io_md": false, 00:05:25.460 "write_zeroes": true, 00:05:25.460 "zcopy": true, 00:05:25.460 "get_zone_info": false, 00:05:25.460 "zone_management": false, 00:05:25.460 "zone_append": false, 00:05:25.460 "compare": false, 00:05:25.460 "compare_and_write": false, 00:05:25.460 "abort": true, 00:05:25.460 "seek_hole": false, 00:05:25.460 "seek_data": false, 00:05:25.460 "copy": true, 00:05:25.460 "nvme_iov_md": false 00:05:25.460 }, 00:05:25.460 "memory_domains": [ 00:05:25.460 { 00:05:25.460 "dma_device_id": "system", 00:05:25.460 "dma_device_type": 1 00:05:25.460 }, 00:05:25.460 { 00:05:25.460 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:25.460 "dma_device_type": 2 00:05:25.460 } 00:05:25.460 ], 00:05:25.460 "driver_specific": {} 00:05:25.460 }, 00:05:25.460 { 00:05:25.460 "name": "Passthru0", 00:05:25.460 "aliases": [ 00:05:25.460 "07c9165c-3285-557f-83ad-5cc1ed2868ca" 00:05:25.460 ], 00:05:25.460 "product_name": "passthru", 00:05:25.460 "block_size": 512, 00:05:25.460 "num_blocks": 16384, 00:05:25.460 "uuid": "07c9165c-3285-557f-83ad-5cc1ed2868ca", 00:05:25.460 "assigned_rate_limits": { 00:05:25.460 "rw_ios_per_sec": 0, 00:05:25.460 "rw_mbytes_per_sec": 0, 00:05:25.460 "r_mbytes_per_sec": 0, 00:05:25.460 "w_mbytes_per_sec": 0 00:05:25.460 }, 00:05:25.460 "claimed": false, 00:05:25.460 "zoned": false, 00:05:25.460 "supported_io_types": { 00:05:25.460 "read": true, 00:05:25.460 "write": true, 00:05:25.460 "unmap": true, 00:05:25.460 "flush": true, 00:05:25.460 "reset": true, 00:05:25.460 "nvme_admin": false, 00:05:25.460 "nvme_io": false, 00:05:25.460 "nvme_io_md": false, 00:05:25.460 "write_zeroes": true, 00:05:25.460 "zcopy": true, 00:05:25.460 "get_zone_info": false, 00:05:25.460 "zone_management": false, 00:05:25.460 "zone_append": false, 00:05:25.460 "compare": false, 00:05:25.460 "compare_and_write": false, 00:05:25.460 "abort": true, 00:05:25.460 "seek_hole": false, 00:05:25.460 "seek_data": false, 00:05:25.460 "copy": true, 00:05:25.460 "nvme_iov_md": false 00:05:25.460 }, 00:05:25.460 "memory_domains": [ 00:05:25.460 { 00:05:25.460 "dma_device_id": "system", 00:05:25.460 "dma_device_type": 1 00:05:25.460 }, 00:05:25.461 { 00:05:25.461 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:25.461 "dma_device_type": 2 00:05:25.461 } 00:05:25.461 ], 00:05:25.461 "driver_specific": { 00:05:25.461 "passthru": { 00:05:25.461 "name": "Passthru0", 00:05:25.461 "base_bdev_name": "Malloc0" 00:05:25.461 } 00:05:25.461 } 00:05:25.461 } 00:05:25.461 ]' 00:05:25.461 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:25.461 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:25.461 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:25.461 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.461 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.461 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.461 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:05:25.461 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.461 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.461 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.461 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:25.461 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.461 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.461 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.461 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:25.461 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:25.461 05:32:58 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:25.461 00:05:25.461 real 0m0.308s 00:05:25.461 user 0m0.180s 00:05:25.461 sys 0m0.060s 00:05:25.461 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:25.461 05:32:58 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.461 ************************************ 00:05:25.461 END TEST rpc_integrity 00:05:25.461 ************************************ 00:05:25.721 05:32:58 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:05:25.721 05:32:58 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:25.721 05:32:58 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:25.721 05:32:58 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:25.721 ************************************ 00:05:25.721 START TEST rpc_plugins 00:05:25.721 ************************************ 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@1129 -- # rpc_plugins 00:05:25.721 05:32:58 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.721 05:32:58 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:05:25.721 05:32:58 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.721 05:32:58 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:05:25.721 { 00:05:25.721 "name": "Malloc1", 00:05:25.721 "aliases": [ 00:05:25.721 "1f2d3c18-5849-448e-8a63-4a6999166a16" 00:05:25.721 ], 00:05:25.721 "product_name": "Malloc disk", 00:05:25.721 "block_size": 4096, 00:05:25.721 "num_blocks": 256, 00:05:25.721 "uuid": "1f2d3c18-5849-448e-8a63-4a6999166a16", 00:05:25.721 "assigned_rate_limits": { 00:05:25.721 "rw_ios_per_sec": 0, 00:05:25.721 "rw_mbytes_per_sec": 0, 00:05:25.721 "r_mbytes_per_sec": 0, 00:05:25.721 "w_mbytes_per_sec": 0 00:05:25.721 }, 00:05:25.721 "claimed": false, 00:05:25.721 "zoned": false, 00:05:25.721 "supported_io_types": { 00:05:25.721 "read": true, 00:05:25.721 "write": true, 00:05:25.721 "unmap": true, 00:05:25.721 "flush": true, 00:05:25.721 "reset": true, 00:05:25.721 "nvme_admin": false, 00:05:25.721 "nvme_io": false, 00:05:25.721 "nvme_io_md": false, 00:05:25.721 "write_zeroes": true, 00:05:25.721 "zcopy": true, 00:05:25.721 "get_zone_info": false, 00:05:25.721 "zone_management": false, 00:05:25.721 "zone_append": false, 00:05:25.721 "compare": false, 00:05:25.721 "compare_and_write": false, 00:05:25.721 "abort": true, 00:05:25.721 "seek_hole": false, 00:05:25.721 "seek_data": false, 00:05:25.721 "copy": true, 00:05:25.721 "nvme_iov_md": false 00:05:25.721 }, 00:05:25.721 "memory_domains": [ 00:05:25.721 { 00:05:25.721 "dma_device_id": "system", 00:05:25.721 "dma_device_type": 1 00:05:25.721 }, 00:05:25.721 { 00:05:25.721 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:25.721 "dma_device_type": 2 00:05:25.721 } 00:05:25.721 ], 00:05:25.721 "driver_specific": {} 00:05:25.721 } 00:05:25.721 ]' 00:05:25.721 05:32:58 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:05:25.721 05:32:58 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:05:25.721 05:32:58 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.721 05:32:58 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:25.721 05:32:58 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.721 05:32:58 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:05:25.721 05:32:58 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:05:25.721 05:32:59 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:05:25.721 00:05:25.721 real 0m0.167s 00:05:25.721 user 0m0.099s 00:05:25.721 sys 0m0.024s 00:05:25.721 05:32:59 rpc.rpc_plugins -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:25.721 05:32:59 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:25.721 ************************************ 00:05:25.721 END TEST rpc_plugins 00:05:25.721 ************************************ 00:05:25.981 05:32:59 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:05:25.981 05:32:59 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:25.981 05:32:59 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:25.981 05:32:59 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:25.981 ************************************ 00:05:25.981 START TEST rpc_trace_cmd_test 00:05:25.981 ************************************ 00:05:25.981 05:32:59 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1129 -- # rpc_trace_cmd_test 00:05:25.981 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:05:25.981 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:05:25.982 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid68835", 00:05:25.982 "tpoint_group_mask": "0x8", 00:05:25.982 "iscsi_conn": { 00:05:25.982 "mask": "0x2", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "scsi": { 00:05:25.982 "mask": "0x4", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "bdev": { 00:05:25.982 "mask": "0x8", 00:05:25.982 "tpoint_mask": "0xffffffffffffffff" 00:05:25.982 }, 00:05:25.982 "nvmf_rdma": { 00:05:25.982 "mask": "0x10", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "nvmf_tcp": { 00:05:25.982 "mask": "0x20", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "ftl": { 00:05:25.982 "mask": "0x40", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "blobfs": { 00:05:25.982 "mask": "0x80", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "dsa": { 00:05:25.982 "mask": "0x200", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "thread": { 00:05:25.982 "mask": "0x400", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "nvme_pcie": { 00:05:25.982 "mask": "0x800", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "iaa": { 00:05:25.982 "mask": "0x1000", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "nvme_tcp": { 00:05:25.982 "mask": "0x2000", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "bdev_nvme": { 00:05:25.982 "mask": "0x4000", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "sock": { 00:05:25.982 "mask": "0x8000", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "blob": { 00:05:25.982 "mask": "0x10000", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "bdev_raid": { 00:05:25.982 "mask": "0x20000", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 }, 00:05:25.982 "scheduler": { 00:05:25.982 "mask": "0x40000", 00:05:25.982 "tpoint_mask": "0x0" 00:05:25.982 } 00:05:25.982 }' 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:05:25.982 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:05:26.242 05:32:59 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:05:26.242 00:05:26.242 real 0m0.251s 00:05:26.242 user 0m0.203s 00:05:26.242 sys 0m0.037s 00:05:26.242 05:32:59 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:26.242 05:32:59 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:26.242 ************************************ 00:05:26.242 END TEST rpc_trace_cmd_test 00:05:26.242 ************************************ 00:05:26.242 05:32:59 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:05:26.242 05:32:59 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:05:26.242 05:32:59 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:05:26.242 05:32:59 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:26.242 05:32:59 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:26.242 05:32:59 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:26.242 ************************************ 00:05:26.242 START TEST rpc_daemon_integrity 00:05:26.242 ************************************ 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:26.242 { 00:05:26.242 "name": "Malloc2", 00:05:26.242 "aliases": [ 00:05:26.242 "916cf859-1e26-470b-ab83-647eb8b729bc" 00:05:26.242 ], 00:05:26.242 "product_name": "Malloc disk", 00:05:26.242 "block_size": 512, 00:05:26.242 "num_blocks": 16384, 00:05:26.242 "uuid": "916cf859-1e26-470b-ab83-647eb8b729bc", 00:05:26.242 "assigned_rate_limits": { 00:05:26.242 "rw_ios_per_sec": 0, 00:05:26.242 "rw_mbytes_per_sec": 0, 00:05:26.242 "r_mbytes_per_sec": 0, 00:05:26.242 "w_mbytes_per_sec": 0 00:05:26.242 }, 00:05:26.242 "claimed": false, 00:05:26.242 "zoned": false, 00:05:26.242 "supported_io_types": { 00:05:26.242 "read": true, 00:05:26.242 "write": true, 00:05:26.242 "unmap": true, 00:05:26.242 "flush": true, 00:05:26.242 "reset": true, 00:05:26.242 "nvme_admin": false, 00:05:26.242 "nvme_io": false, 00:05:26.242 "nvme_io_md": false, 00:05:26.242 "write_zeroes": true, 00:05:26.242 "zcopy": true, 00:05:26.242 "get_zone_info": false, 00:05:26.242 "zone_management": false, 00:05:26.242 "zone_append": false, 00:05:26.242 "compare": false, 00:05:26.242 "compare_and_write": false, 00:05:26.242 "abort": true, 00:05:26.242 "seek_hole": false, 00:05:26.242 "seek_data": false, 00:05:26.242 "copy": true, 00:05:26.242 "nvme_iov_md": false 00:05:26.242 }, 00:05:26.242 "memory_domains": [ 00:05:26.242 { 00:05:26.242 "dma_device_id": "system", 00:05:26.242 "dma_device_type": 1 00:05:26.242 }, 00:05:26.242 { 00:05:26.242 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:26.242 "dma_device_type": 2 00:05:26.242 } 00:05:26.242 ], 00:05:26.242 "driver_specific": {} 00:05:26.242 } 00:05:26.242 ]' 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.242 [2024-12-07 05:32:59.573480] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:05:26.242 [2024-12-07 05:32:59.573575] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:26.242 [2024-12-07 05:32:59.573602] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:05:26.242 [2024-12-07 05:32:59.573612] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:26.242 [2024-12-07 05:32:59.576198] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:26.242 [2024-12-07 05:32:59.576240] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:26.242 Passthru0 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:26.242 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.503 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:26.503 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:26.503 { 00:05:26.503 "name": "Malloc2", 00:05:26.503 "aliases": [ 00:05:26.503 "916cf859-1e26-470b-ab83-647eb8b729bc" 00:05:26.503 ], 00:05:26.503 "product_name": "Malloc disk", 00:05:26.503 "block_size": 512, 00:05:26.503 "num_blocks": 16384, 00:05:26.503 "uuid": "916cf859-1e26-470b-ab83-647eb8b729bc", 00:05:26.503 "assigned_rate_limits": { 00:05:26.503 "rw_ios_per_sec": 0, 00:05:26.503 "rw_mbytes_per_sec": 0, 00:05:26.503 "r_mbytes_per_sec": 0, 00:05:26.503 "w_mbytes_per_sec": 0 00:05:26.503 }, 00:05:26.503 "claimed": true, 00:05:26.503 "claim_type": "exclusive_write", 00:05:26.503 "zoned": false, 00:05:26.503 "supported_io_types": { 00:05:26.503 "read": true, 00:05:26.503 "write": true, 00:05:26.503 "unmap": true, 00:05:26.503 "flush": true, 00:05:26.503 "reset": true, 00:05:26.503 "nvme_admin": false, 00:05:26.503 "nvme_io": false, 00:05:26.503 "nvme_io_md": false, 00:05:26.503 "write_zeroes": true, 00:05:26.503 "zcopy": true, 00:05:26.503 "get_zone_info": false, 00:05:26.503 "zone_management": false, 00:05:26.503 "zone_append": false, 00:05:26.503 "compare": false, 00:05:26.503 "compare_and_write": false, 00:05:26.503 "abort": true, 00:05:26.503 "seek_hole": false, 00:05:26.503 "seek_data": false, 00:05:26.503 "copy": true, 00:05:26.503 "nvme_iov_md": false 00:05:26.503 }, 00:05:26.503 "memory_domains": [ 00:05:26.503 { 00:05:26.503 "dma_device_id": "system", 00:05:26.503 "dma_device_type": 1 00:05:26.503 }, 00:05:26.503 { 00:05:26.503 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:26.503 "dma_device_type": 2 00:05:26.503 } 00:05:26.503 ], 00:05:26.503 "driver_specific": {} 00:05:26.503 }, 00:05:26.503 { 00:05:26.503 "name": "Passthru0", 00:05:26.503 "aliases": [ 00:05:26.503 "ee5d9437-0be9-5b6b-903b-84ac307e8bac" 00:05:26.503 ], 00:05:26.503 "product_name": "passthru", 00:05:26.503 "block_size": 512, 00:05:26.503 "num_blocks": 16384, 00:05:26.503 "uuid": "ee5d9437-0be9-5b6b-903b-84ac307e8bac", 00:05:26.503 "assigned_rate_limits": { 00:05:26.503 "rw_ios_per_sec": 0, 00:05:26.503 "rw_mbytes_per_sec": 0, 00:05:26.503 "r_mbytes_per_sec": 0, 00:05:26.503 "w_mbytes_per_sec": 0 00:05:26.503 }, 00:05:26.503 "claimed": false, 00:05:26.503 "zoned": false, 00:05:26.503 "supported_io_types": { 00:05:26.503 "read": true, 00:05:26.503 "write": true, 00:05:26.503 "unmap": true, 00:05:26.503 "flush": true, 00:05:26.503 "reset": true, 00:05:26.503 "nvme_admin": false, 00:05:26.503 "nvme_io": false, 00:05:26.503 "nvme_io_md": false, 00:05:26.503 "write_zeroes": true, 00:05:26.503 "zcopy": true, 00:05:26.503 "get_zone_info": false, 00:05:26.503 "zone_management": false, 00:05:26.503 "zone_append": false, 00:05:26.503 "compare": false, 00:05:26.503 "compare_and_write": false, 00:05:26.503 "abort": true, 00:05:26.503 "seek_hole": false, 00:05:26.503 "seek_data": false, 00:05:26.503 "copy": true, 00:05:26.503 "nvme_iov_md": false 00:05:26.503 }, 00:05:26.503 "memory_domains": [ 00:05:26.503 { 00:05:26.503 "dma_device_id": "system", 00:05:26.503 "dma_device_type": 1 00:05:26.503 }, 00:05:26.503 { 00:05:26.503 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:26.503 "dma_device_type": 2 00:05:26.503 } 00:05:26.503 ], 00:05:26.503 "driver_specific": { 00:05:26.503 "passthru": { 00:05:26.503 "name": "Passthru0", 00:05:26.504 "base_bdev_name": "Malloc2" 00:05:26.504 } 00:05:26.504 } 00:05:26.504 } 00:05:26.504 ]' 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:26.504 00:05:26.504 real 0m0.308s 00:05:26.504 user 0m0.173s 00:05:26.504 sys 0m0.065s 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:26.504 05:32:59 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.504 ************************************ 00:05:26.504 END TEST rpc_daemon_integrity 00:05:26.504 ************************************ 00:05:26.504 05:32:59 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:05:26.504 05:32:59 rpc -- rpc/rpc.sh@84 -- # killprocess 68835 00:05:26.504 05:32:59 rpc -- common/autotest_common.sh@954 -- # '[' -z 68835 ']' 00:05:26.504 05:32:59 rpc -- common/autotest_common.sh@958 -- # kill -0 68835 00:05:26.504 05:32:59 rpc -- common/autotest_common.sh@959 -- # uname 00:05:26.504 05:32:59 rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:26.504 05:32:59 rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 68835 00:05:26.504 05:32:59 rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:26.504 05:32:59 rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:26.504 killing process with pid 68835 00:05:26.504 05:32:59 rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 68835' 00:05:26.504 05:32:59 rpc -- common/autotest_common.sh@973 -- # kill 68835 00:05:26.504 05:32:59 rpc -- common/autotest_common.sh@978 -- # wait 68835 00:05:27.075 00:05:27.075 real 0m2.853s 00:05:27.075 user 0m3.450s 00:05:27.075 sys 0m0.871s 00:05:27.075 05:33:00 rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:27.075 05:33:00 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:27.075 ************************************ 00:05:27.075 END TEST rpc 00:05:27.075 ************************************ 00:05:27.075 05:33:00 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:27.075 05:33:00 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:27.075 05:33:00 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:27.075 05:33:00 -- common/autotest_common.sh@10 -- # set +x 00:05:27.075 ************************************ 00:05:27.075 START TEST skip_rpc 00:05:27.075 ************************************ 00:05:27.075 05:33:00 skip_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:27.075 * Looking for test storage... 00:05:27.075 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:27.076 05:33:00 skip_rpc -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:27.076 05:33:00 skip_rpc -- common/autotest_common.sh@1711 -- # lcov --version 00:05:27.076 05:33:00 skip_rpc -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:27.336 05:33:00 skip_rpc -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@345 -- # : 1 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:27.336 05:33:00 skip_rpc -- scripts/common.sh@368 -- # return 0 00:05:27.336 05:33:00 skip_rpc -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:27.336 05:33:00 skip_rpc -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:27.336 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:27.336 --rc genhtml_branch_coverage=1 00:05:27.336 --rc genhtml_function_coverage=1 00:05:27.336 --rc genhtml_legend=1 00:05:27.336 --rc geninfo_all_blocks=1 00:05:27.336 --rc geninfo_unexecuted_blocks=1 00:05:27.336 00:05:27.336 ' 00:05:27.336 05:33:00 skip_rpc -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:27.336 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:27.336 --rc genhtml_branch_coverage=1 00:05:27.336 --rc genhtml_function_coverage=1 00:05:27.336 --rc genhtml_legend=1 00:05:27.336 --rc geninfo_all_blocks=1 00:05:27.336 --rc geninfo_unexecuted_blocks=1 00:05:27.336 00:05:27.336 ' 00:05:27.336 05:33:00 skip_rpc -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:27.336 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:27.336 --rc genhtml_branch_coverage=1 00:05:27.336 --rc genhtml_function_coverage=1 00:05:27.336 --rc genhtml_legend=1 00:05:27.336 --rc geninfo_all_blocks=1 00:05:27.336 --rc geninfo_unexecuted_blocks=1 00:05:27.336 00:05:27.336 ' 00:05:27.336 05:33:00 skip_rpc -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:27.336 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:27.336 --rc genhtml_branch_coverage=1 00:05:27.336 --rc genhtml_function_coverage=1 00:05:27.336 --rc genhtml_legend=1 00:05:27.336 --rc geninfo_all_blocks=1 00:05:27.336 --rc geninfo_unexecuted_blocks=1 00:05:27.336 00:05:27.336 ' 00:05:27.336 05:33:00 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:27.336 05:33:00 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:27.336 05:33:00 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:05:27.336 05:33:00 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:27.336 05:33:00 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:27.336 05:33:00 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:27.336 ************************************ 00:05:27.336 START TEST skip_rpc 00:05:27.336 ************************************ 00:05:27.336 05:33:00 skip_rpc.skip_rpc -- common/autotest_common.sh@1129 -- # test_skip_rpc 00:05:27.336 05:33:00 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=69037 00:05:27.336 05:33:00 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:05:27.336 05:33:00 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:27.336 05:33:00 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:05:27.336 [2024-12-07 05:33:00.582918] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:27.336 [2024-12-07 05:33:00.583068] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69037 ] 00:05:27.596 [2024-12-07 05:33:00.739682] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:27.596 [2024-12-07 05:33:00.768866] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:32.886 05:33:05 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:05:32.886 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # local es=0 00:05:32.886 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd spdk_get_version 00:05:32.886 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # rpc_cmd spdk_get_version 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # es=1 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 69037 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # '[' -z 69037 ']' 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # kill -0 69037 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # uname 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69037 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69037' 00:05:32.887 killing process with pid 69037 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@973 -- # kill 69037 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@978 -- # wait 69037 00:05:32.887 00:05:32.887 real 0m5.424s 00:05:32.887 user 0m5.039s 00:05:32.887 sys 0m0.313s 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:32.887 05:33:05 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:32.887 ************************************ 00:05:32.887 END TEST skip_rpc 00:05:32.887 ************************************ 00:05:32.887 05:33:05 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:05:32.887 05:33:05 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:32.887 05:33:05 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:32.887 05:33:05 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:32.887 ************************************ 00:05:32.887 START TEST skip_rpc_with_json 00:05:32.887 ************************************ 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_json 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=69124 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 69124 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # '[' -z 69124 ']' 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:32.887 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:32.887 05:33:05 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:32.887 [2024-12-07 05:33:06.062302] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:32.887 [2024-12-07 05:33:06.062437] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69124 ] 00:05:32.887 [2024-12-07 05:33:06.218399] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:32.887 [2024-12-07 05:33:06.244445] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@868 -- # return 0 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:33.839 [2024-12-07 05:33:06.909763] nvmf_rpc.c:2707:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:05:33.839 request: 00:05:33.839 { 00:05:33.839 "trtype": "tcp", 00:05:33.839 "method": "nvmf_get_transports", 00:05:33.839 "req_id": 1 00:05:33.839 } 00:05:33.839 Got JSON-RPC error response 00:05:33.839 response: 00:05:33.839 { 00:05:33.839 "code": -19, 00:05:33.839 "message": "No such device" 00:05:33.839 } 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:33.839 [2024-12-07 05:33:06.921887] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:33.839 05:33:06 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:33.839 05:33:07 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:33.839 05:33:07 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:33.839 { 00:05:33.839 "subsystems": [ 00:05:33.839 { 00:05:33.839 "subsystem": "fsdev", 00:05:33.839 "config": [ 00:05:33.839 { 00:05:33.839 "method": "fsdev_set_opts", 00:05:33.839 "params": { 00:05:33.839 "fsdev_io_pool_size": 65535, 00:05:33.839 "fsdev_io_cache_size": 256 00:05:33.839 } 00:05:33.839 } 00:05:33.839 ] 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "subsystem": "keyring", 00:05:33.839 "config": [] 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "subsystem": "iobuf", 00:05:33.839 "config": [ 00:05:33.839 { 00:05:33.839 "method": "iobuf_set_options", 00:05:33.839 "params": { 00:05:33.839 "small_pool_count": 8192, 00:05:33.839 "large_pool_count": 1024, 00:05:33.839 "small_bufsize": 8192, 00:05:33.839 "large_bufsize": 135168, 00:05:33.839 "enable_numa": false 00:05:33.839 } 00:05:33.839 } 00:05:33.839 ] 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "subsystem": "sock", 00:05:33.839 "config": [ 00:05:33.839 { 00:05:33.839 "method": "sock_set_default_impl", 00:05:33.839 "params": { 00:05:33.839 "impl_name": "posix" 00:05:33.839 } 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "method": "sock_impl_set_options", 00:05:33.839 "params": { 00:05:33.839 "impl_name": "ssl", 00:05:33.839 "recv_buf_size": 4096, 00:05:33.839 "send_buf_size": 4096, 00:05:33.839 "enable_recv_pipe": true, 00:05:33.839 "enable_quickack": false, 00:05:33.839 "enable_placement_id": 0, 00:05:33.839 "enable_zerocopy_send_server": true, 00:05:33.839 "enable_zerocopy_send_client": false, 00:05:33.839 "zerocopy_threshold": 0, 00:05:33.839 "tls_version": 0, 00:05:33.839 "enable_ktls": false 00:05:33.839 } 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "method": "sock_impl_set_options", 00:05:33.839 "params": { 00:05:33.839 "impl_name": "posix", 00:05:33.839 "recv_buf_size": 2097152, 00:05:33.839 "send_buf_size": 2097152, 00:05:33.839 "enable_recv_pipe": true, 00:05:33.839 "enable_quickack": false, 00:05:33.839 "enable_placement_id": 0, 00:05:33.839 "enable_zerocopy_send_server": true, 00:05:33.839 "enable_zerocopy_send_client": false, 00:05:33.839 "zerocopy_threshold": 0, 00:05:33.839 "tls_version": 0, 00:05:33.839 "enable_ktls": false 00:05:33.839 } 00:05:33.839 } 00:05:33.839 ] 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "subsystem": "vmd", 00:05:33.839 "config": [] 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "subsystem": "accel", 00:05:33.839 "config": [ 00:05:33.839 { 00:05:33.839 "method": "accel_set_options", 00:05:33.839 "params": { 00:05:33.839 "small_cache_size": 128, 00:05:33.839 "large_cache_size": 16, 00:05:33.839 "task_count": 2048, 00:05:33.839 "sequence_count": 2048, 00:05:33.839 "buf_count": 2048 00:05:33.839 } 00:05:33.839 } 00:05:33.839 ] 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "subsystem": "bdev", 00:05:33.839 "config": [ 00:05:33.839 { 00:05:33.839 "method": "bdev_set_options", 00:05:33.839 "params": { 00:05:33.839 "bdev_io_pool_size": 65535, 00:05:33.839 "bdev_io_cache_size": 256, 00:05:33.839 "bdev_auto_examine": true, 00:05:33.839 "iobuf_small_cache_size": 128, 00:05:33.839 "iobuf_large_cache_size": 16 00:05:33.839 } 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "method": "bdev_raid_set_options", 00:05:33.839 "params": { 00:05:33.839 "process_window_size_kb": 1024, 00:05:33.839 "process_max_bandwidth_mb_sec": 0 00:05:33.839 } 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "method": "bdev_iscsi_set_options", 00:05:33.839 "params": { 00:05:33.839 "timeout_sec": 30 00:05:33.839 } 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "method": "bdev_nvme_set_options", 00:05:33.839 "params": { 00:05:33.839 "action_on_timeout": "none", 00:05:33.839 "timeout_us": 0, 00:05:33.839 "timeout_admin_us": 0, 00:05:33.839 "keep_alive_timeout_ms": 10000, 00:05:33.839 "arbitration_burst": 0, 00:05:33.839 "low_priority_weight": 0, 00:05:33.839 "medium_priority_weight": 0, 00:05:33.839 "high_priority_weight": 0, 00:05:33.839 "nvme_adminq_poll_period_us": 10000, 00:05:33.839 "nvme_ioq_poll_period_us": 0, 00:05:33.839 "io_queue_requests": 0, 00:05:33.839 "delay_cmd_submit": true, 00:05:33.839 "transport_retry_count": 4, 00:05:33.839 "bdev_retry_count": 3, 00:05:33.839 "transport_ack_timeout": 0, 00:05:33.839 "ctrlr_loss_timeout_sec": 0, 00:05:33.839 "reconnect_delay_sec": 0, 00:05:33.839 "fast_io_fail_timeout_sec": 0, 00:05:33.839 "disable_auto_failback": false, 00:05:33.839 "generate_uuids": false, 00:05:33.839 "transport_tos": 0, 00:05:33.839 "nvme_error_stat": false, 00:05:33.839 "rdma_srq_size": 0, 00:05:33.839 "io_path_stat": false, 00:05:33.839 "allow_accel_sequence": false, 00:05:33.839 "rdma_max_cq_size": 0, 00:05:33.839 "rdma_cm_event_timeout_ms": 0, 00:05:33.839 "dhchap_digests": [ 00:05:33.839 "sha256", 00:05:33.839 "sha384", 00:05:33.839 "sha512" 00:05:33.839 ], 00:05:33.839 "dhchap_dhgroups": [ 00:05:33.839 "null", 00:05:33.839 "ffdhe2048", 00:05:33.839 "ffdhe3072", 00:05:33.839 "ffdhe4096", 00:05:33.839 "ffdhe6144", 00:05:33.839 "ffdhe8192" 00:05:33.839 ] 00:05:33.839 } 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "method": "bdev_nvme_set_hotplug", 00:05:33.839 "params": { 00:05:33.839 "period_us": 100000, 00:05:33.839 "enable": false 00:05:33.839 } 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "method": "bdev_wait_for_examine" 00:05:33.839 } 00:05:33.839 ] 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "subsystem": "scsi", 00:05:33.839 "config": null 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "subsystem": "scheduler", 00:05:33.839 "config": [ 00:05:33.839 { 00:05:33.839 "method": "framework_set_scheduler", 00:05:33.839 "params": { 00:05:33.839 "name": "static" 00:05:33.839 } 00:05:33.839 } 00:05:33.839 ] 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "subsystem": "vhost_scsi", 00:05:33.839 "config": [] 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "subsystem": "vhost_blk", 00:05:33.839 "config": [] 00:05:33.839 }, 00:05:33.839 { 00:05:33.839 "subsystem": "ublk", 00:05:33.839 "config": [] 00:05:33.839 }, 00:05:33.839 { 00:05:33.840 "subsystem": "nbd", 00:05:33.840 "config": [] 00:05:33.840 }, 00:05:33.840 { 00:05:33.840 "subsystem": "nvmf", 00:05:33.840 "config": [ 00:05:33.840 { 00:05:33.840 "method": "nvmf_set_config", 00:05:33.840 "params": { 00:05:33.840 "discovery_filter": "match_any", 00:05:33.840 "admin_cmd_passthru": { 00:05:33.840 "identify_ctrlr": false 00:05:33.840 }, 00:05:33.840 "dhchap_digests": [ 00:05:33.840 "sha256", 00:05:33.840 "sha384", 00:05:33.840 "sha512" 00:05:33.840 ], 00:05:33.840 "dhchap_dhgroups": [ 00:05:33.840 "null", 00:05:33.840 "ffdhe2048", 00:05:33.840 "ffdhe3072", 00:05:33.840 "ffdhe4096", 00:05:33.840 "ffdhe6144", 00:05:33.840 "ffdhe8192" 00:05:33.840 ] 00:05:33.840 } 00:05:33.840 }, 00:05:33.840 { 00:05:33.840 "method": "nvmf_set_max_subsystems", 00:05:33.840 "params": { 00:05:33.840 "max_subsystems": 1024 00:05:33.840 } 00:05:33.840 }, 00:05:33.840 { 00:05:33.840 "method": "nvmf_set_crdt", 00:05:33.840 "params": { 00:05:33.840 "crdt1": 0, 00:05:33.840 "crdt2": 0, 00:05:33.840 "crdt3": 0 00:05:33.840 } 00:05:33.840 }, 00:05:33.840 { 00:05:33.840 "method": "nvmf_create_transport", 00:05:33.840 "params": { 00:05:33.840 "trtype": "TCP", 00:05:33.840 "max_queue_depth": 128, 00:05:33.840 "max_io_qpairs_per_ctrlr": 127, 00:05:33.840 "in_capsule_data_size": 4096, 00:05:33.840 "max_io_size": 131072, 00:05:33.840 "io_unit_size": 131072, 00:05:33.840 "max_aq_depth": 128, 00:05:33.840 "num_shared_buffers": 511, 00:05:33.840 "buf_cache_size": 4294967295, 00:05:33.840 "dif_insert_or_strip": false, 00:05:33.840 "zcopy": false, 00:05:33.840 "c2h_success": true, 00:05:33.840 "sock_priority": 0, 00:05:33.840 "abort_timeout_sec": 1, 00:05:33.840 "ack_timeout": 0, 00:05:33.840 "data_wr_pool_size": 0 00:05:33.840 } 00:05:33.840 } 00:05:33.840 ] 00:05:33.840 }, 00:05:33.840 { 00:05:33.840 "subsystem": "iscsi", 00:05:33.840 "config": [ 00:05:33.840 { 00:05:33.840 "method": "iscsi_set_options", 00:05:33.840 "params": { 00:05:33.840 "node_base": "iqn.2016-06.io.spdk", 00:05:33.840 "max_sessions": 128, 00:05:33.840 "max_connections_per_session": 2, 00:05:33.840 "max_queue_depth": 64, 00:05:33.840 "default_time2wait": 2, 00:05:33.840 "default_time2retain": 20, 00:05:33.840 "first_burst_length": 8192, 00:05:33.840 "immediate_data": true, 00:05:33.840 "allow_duplicated_isid": false, 00:05:33.840 "error_recovery_level": 0, 00:05:33.840 "nop_timeout": 60, 00:05:33.840 "nop_in_interval": 30, 00:05:33.840 "disable_chap": false, 00:05:33.840 "require_chap": false, 00:05:33.840 "mutual_chap": false, 00:05:33.840 "chap_group": 0, 00:05:33.840 "max_large_datain_per_connection": 64, 00:05:33.840 "max_r2t_per_connection": 4, 00:05:33.840 "pdu_pool_size": 36864, 00:05:33.840 "immediate_data_pool_size": 16384, 00:05:33.840 "data_out_pool_size": 2048 00:05:33.840 } 00:05:33.840 } 00:05:33.840 ] 00:05:33.840 } 00:05:33.840 ] 00:05:33.840 } 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 69124 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 69124 ']' 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 69124 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69124 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:33.840 killing process with pid 69124 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69124' 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 69124 00:05:33.840 05:33:07 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 69124 00:05:34.411 05:33:07 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=69153 00:05:34.411 05:33:07 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:34.411 05:33:07 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 69153 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 69153 ']' 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 69153 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69153 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:39.693 killing process with pid 69153 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69153' 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 69153 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 69153 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:39.693 00:05:39.693 real 0m6.957s 00:05:39.693 user 0m6.565s 00:05:39.693 sys 0m0.706s 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:39.693 ************************************ 00:05:39.693 END TEST skip_rpc_with_json 00:05:39.693 ************************************ 00:05:39.693 05:33:12 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:05:39.693 05:33:12 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:39.693 05:33:12 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:39.693 05:33:12 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:39.693 ************************************ 00:05:39.693 START TEST skip_rpc_with_delay 00:05:39.693 ************************************ 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_delay 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # local es=0 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:39.693 05:33:12 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:39.693 05:33:13 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:39.693 05:33:13 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:05:39.693 05:33:13 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:39.953 [2024-12-07 05:33:13.095153] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:05:39.953 05:33:13 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # es=1 00:05:39.953 05:33:13 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:39.953 05:33:13 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:39.953 05:33:13 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:39.953 00:05:39.953 real 0m0.166s 00:05:39.953 user 0m0.087s 00:05:39.953 sys 0m0.077s 00:05:39.953 05:33:13 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:39.953 05:33:13 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:05:39.953 ************************************ 00:05:39.953 END TEST skip_rpc_with_delay 00:05:39.953 ************************************ 00:05:39.953 05:33:13 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:05:39.953 05:33:13 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:05:39.953 05:33:13 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:05:39.953 05:33:13 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:39.953 05:33:13 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:39.953 05:33:13 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:39.953 ************************************ 00:05:39.953 START TEST exit_on_failed_rpc_init 00:05:39.953 ************************************ 00:05:39.953 05:33:13 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1129 -- # test_exit_on_failed_rpc_init 00:05:39.953 05:33:13 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=69264 00:05:39.953 05:33:13 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:39.953 05:33:13 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 69264 00:05:39.954 05:33:13 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # '[' -z 69264 ']' 00:05:39.954 05:33:13 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:39.954 05:33:13 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:39.954 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:39.954 05:33:13 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:39.954 05:33:13 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:39.954 05:33:13 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:05:40.213 [2024-12-07 05:33:13.328437] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:40.213 [2024-12-07 05:33:13.328566] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69264 ] 00:05:40.213 [2024-12-07 05:33:13.486450] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:40.213 [2024-12-07 05:33:13.514319] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@868 -- # return 0 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # local es=0 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:05:41.153 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:41.153 [2024-12-07 05:33:14.280277] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:41.154 [2024-12-07 05:33:14.280411] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69277 ] 00:05:41.154 [2024-12-07 05:33:14.434007] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:41.154 [2024-12-07 05:33:14.461971] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:41.154 [2024-12-07 05:33:14.462064] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:05:41.154 [2024-12-07 05:33:14.462083] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:05:41.154 [2024-12-07 05:33:14.462099] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:05:41.413 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # es=234 00:05:41.413 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:41.413 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@664 -- # es=106 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@665 -- # case "$es" in 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@672 -- # es=1 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 69264 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # '[' -z 69264 ']' 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # kill -0 69264 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # uname 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69264 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:41.414 killing process with pid 69264 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69264' 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@973 -- # kill 69264 00:05:41.414 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@978 -- # wait 69264 00:05:41.674 00:05:41.674 real 0m1.719s 00:05:41.674 user 0m1.862s 00:05:41.674 sys 0m0.486s 00:05:41.674 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:41.674 05:33:14 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:05:41.674 ************************************ 00:05:41.674 END TEST exit_on_failed_rpc_init 00:05:41.674 ************************************ 00:05:41.674 05:33:15 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:41.674 00:05:41.674 real 0m14.761s 00:05:41.674 user 0m13.758s 00:05:41.674 sys 0m1.895s 00:05:41.674 ************************************ 00:05:41.674 END TEST skip_rpc 00:05:41.674 ************************************ 00:05:41.674 05:33:15 skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:41.674 05:33:15 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:41.933 05:33:15 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:05:41.933 05:33:15 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:41.933 05:33:15 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:41.933 05:33:15 -- common/autotest_common.sh@10 -- # set +x 00:05:41.933 ************************************ 00:05:41.933 START TEST rpc_client 00:05:41.933 ************************************ 00:05:41.933 05:33:15 rpc_client -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:05:41.933 * Looking for test storage... 00:05:41.933 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:05:41.933 05:33:15 rpc_client -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:41.933 05:33:15 rpc_client -- common/autotest_common.sh@1711 -- # lcov --version 00:05:41.933 05:33:15 rpc_client -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:41.933 05:33:15 rpc_client -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:41.933 05:33:15 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:41.933 05:33:15 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:41.933 05:33:15 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:41.933 05:33:15 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:05:41.933 05:33:15 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:05:41.933 05:33:15 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:05:41.933 05:33:15 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:05:41.933 05:33:15 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:05:41.933 05:33:15 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@345 -- # : 1 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@353 -- # local d=1 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@355 -- # echo 1 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@353 -- # local d=2 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:41.934 05:33:15 rpc_client -- scripts/common.sh@355 -- # echo 2 00:05:42.193 05:33:15 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:05:42.193 05:33:15 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:42.193 05:33:15 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:42.193 05:33:15 rpc_client -- scripts/common.sh@368 -- # return 0 00:05:42.193 05:33:15 rpc_client -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:42.193 05:33:15 rpc_client -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:42.193 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.193 --rc genhtml_branch_coverage=1 00:05:42.193 --rc genhtml_function_coverage=1 00:05:42.193 --rc genhtml_legend=1 00:05:42.193 --rc geninfo_all_blocks=1 00:05:42.193 --rc geninfo_unexecuted_blocks=1 00:05:42.193 00:05:42.193 ' 00:05:42.193 05:33:15 rpc_client -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:42.193 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.193 --rc genhtml_branch_coverage=1 00:05:42.193 --rc genhtml_function_coverage=1 00:05:42.193 --rc genhtml_legend=1 00:05:42.193 --rc geninfo_all_blocks=1 00:05:42.193 --rc geninfo_unexecuted_blocks=1 00:05:42.193 00:05:42.193 ' 00:05:42.193 05:33:15 rpc_client -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:42.193 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.193 --rc genhtml_branch_coverage=1 00:05:42.193 --rc genhtml_function_coverage=1 00:05:42.193 --rc genhtml_legend=1 00:05:42.193 --rc geninfo_all_blocks=1 00:05:42.193 --rc geninfo_unexecuted_blocks=1 00:05:42.193 00:05:42.193 ' 00:05:42.193 05:33:15 rpc_client -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:42.193 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.193 --rc genhtml_branch_coverage=1 00:05:42.193 --rc genhtml_function_coverage=1 00:05:42.193 --rc genhtml_legend=1 00:05:42.193 --rc geninfo_all_blocks=1 00:05:42.193 --rc geninfo_unexecuted_blocks=1 00:05:42.193 00:05:42.193 ' 00:05:42.193 05:33:15 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:05:42.193 OK 00:05:42.193 05:33:15 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:05:42.193 00:05:42.193 real 0m0.298s 00:05:42.193 user 0m0.151s 00:05:42.193 sys 0m0.163s 00:05:42.193 ************************************ 00:05:42.193 END TEST rpc_client 00:05:42.193 ************************************ 00:05:42.193 05:33:15 rpc_client -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:42.193 05:33:15 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:05:42.193 05:33:15 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:05:42.193 05:33:15 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:42.193 05:33:15 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:42.193 05:33:15 -- common/autotest_common.sh@10 -- # set +x 00:05:42.193 ************************************ 00:05:42.193 START TEST json_config 00:05:42.193 ************************************ 00:05:42.193 05:33:15 json_config -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:05:42.193 05:33:15 json_config -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:42.193 05:33:15 json_config -- common/autotest_common.sh@1711 -- # lcov --version 00:05:42.193 05:33:15 json_config -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:42.453 05:33:15 json_config -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:42.453 05:33:15 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:42.453 05:33:15 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:42.453 05:33:15 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:42.453 05:33:15 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:05:42.453 05:33:15 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:05:42.453 05:33:15 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:05:42.453 05:33:15 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:05:42.453 05:33:15 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:05:42.453 05:33:15 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:05:42.453 05:33:15 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:05:42.453 05:33:15 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:42.453 05:33:15 json_config -- scripts/common.sh@344 -- # case "$op" in 00:05:42.453 05:33:15 json_config -- scripts/common.sh@345 -- # : 1 00:05:42.453 05:33:15 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:42.453 05:33:15 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:42.453 05:33:15 json_config -- scripts/common.sh@365 -- # decimal 1 00:05:42.453 05:33:15 json_config -- scripts/common.sh@353 -- # local d=1 00:05:42.453 05:33:15 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:42.453 05:33:15 json_config -- scripts/common.sh@355 -- # echo 1 00:05:42.453 05:33:15 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:05:42.453 05:33:15 json_config -- scripts/common.sh@366 -- # decimal 2 00:05:42.453 05:33:15 json_config -- scripts/common.sh@353 -- # local d=2 00:05:42.453 05:33:15 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:42.453 05:33:15 json_config -- scripts/common.sh@355 -- # echo 2 00:05:42.453 05:33:15 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:05:42.453 05:33:15 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:42.453 05:33:15 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:42.453 05:33:15 json_config -- scripts/common.sh@368 -- # return 0 00:05:42.453 05:33:15 json_config -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:42.453 05:33:15 json_config -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:42.453 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.453 --rc genhtml_branch_coverage=1 00:05:42.453 --rc genhtml_function_coverage=1 00:05:42.453 --rc genhtml_legend=1 00:05:42.454 --rc geninfo_all_blocks=1 00:05:42.454 --rc geninfo_unexecuted_blocks=1 00:05:42.454 00:05:42.454 ' 00:05:42.454 05:33:15 json_config -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:42.454 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.454 --rc genhtml_branch_coverage=1 00:05:42.454 --rc genhtml_function_coverage=1 00:05:42.454 --rc genhtml_legend=1 00:05:42.454 --rc geninfo_all_blocks=1 00:05:42.454 --rc geninfo_unexecuted_blocks=1 00:05:42.454 00:05:42.454 ' 00:05:42.454 05:33:15 json_config -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:42.454 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.454 --rc genhtml_branch_coverage=1 00:05:42.454 --rc genhtml_function_coverage=1 00:05:42.454 --rc genhtml_legend=1 00:05:42.454 --rc geninfo_all_blocks=1 00:05:42.454 --rc geninfo_unexecuted_blocks=1 00:05:42.454 00:05:42.454 ' 00:05:42.454 05:33:15 json_config -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:42.454 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.454 --rc genhtml_branch_coverage=1 00:05:42.454 --rc genhtml_function_coverage=1 00:05:42.454 --rc genhtml_legend=1 00:05:42.454 --rc geninfo_all_blocks=1 00:05:42.454 --rc geninfo_unexecuted_blocks=1 00:05:42.454 00:05:42.454 ' 00:05:42.454 05:33:15 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@7 -- # uname -s 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:1b6716e3-b0c8-480d-9903-786aaf87e8fc 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=1b6716e3-b0c8-480d-9903-786aaf87e8fc 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:42.454 05:33:15 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:05:42.454 05:33:15 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:42.454 05:33:15 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:42.454 05:33:15 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:42.454 05:33:15 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.454 05:33:15 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.454 05:33:15 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.454 05:33:15 json_config -- paths/export.sh@5 -- # export PATH 00:05:42.454 05:33:15 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@51 -- # : 0 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:42.454 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:42.454 05:33:15 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:42.454 05:33:15 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:05:42.454 05:33:15 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:05:42.454 05:33:15 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:05:42.454 05:33:15 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:05:42.454 05:33:15 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:05:42.454 05:33:15 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:05:42.454 WARNING: No tests are enabled so not running JSON configuration tests 00:05:42.454 05:33:15 json_config -- json_config/json_config.sh@28 -- # exit 0 00:05:42.454 00:05:42.454 real 0m0.221s 00:05:42.454 user 0m0.137s 00:05:42.454 sys 0m0.087s 00:05:42.454 05:33:15 json_config -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:42.454 05:33:15 json_config -- common/autotest_common.sh@10 -- # set +x 00:05:42.454 ************************************ 00:05:42.454 END TEST json_config 00:05:42.454 ************************************ 00:05:42.454 05:33:15 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:05:42.454 05:33:15 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:42.454 05:33:15 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:42.454 05:33:15 -- common/autotest_common.sh@10 -- # set +x 00:05:42.454 ************************************ 00:05:42.454 START TEST json_config_extra_key 00:05:42.454 ************************************ 00:05:42.454 05:33:15 json_config_extra_key -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:05:42.454 05:33:15 json_config_extra_key -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:42.454 05:33:15 json_config_extra_key -- common/autotest_common.sh@1711 -- # lcov --version 00:05:42.454 05:33:15 json_config_extra_key -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:42.716 05:33:15 json_config_extra_key -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:05:42.716 05:33:15 json_config_extra_key -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:42.716 05:33:15 json_config_extra_key -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:42.716 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.716 --rc genhtml_branch_coverage=1 00:05:42.716 --rc genhtml_function_coverage=1 00:05:42.716 --rc genhtml_legend=1 00:05:42.716 --rc geninfo_all_blocks=1 00:05:42.716 --rc geninfo_unexecuted_blocks=1 00:05:42.716 00:05:42.716 ' 00:05:42.716 05:33:15 json_config_extra_key -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:42.716 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.716 --rc genhtml_branch_coverage=1 00:05:42.716 --rc genhtml_function_coverage=1 00:05:42.716 --rc genhtml_legend=1 00:05:42.716 --rc geninfo_all_blocks=1 00:05:42.716 --rc geninfo_unexecuted_blocks=1 00:05:42.716 00:05:42.716 ' 00:05:42.716 05:33:15 json_config_extra_key -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:42.716 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.716 --rc genhtml_branch_coverage=1 00:05:42.716 --rc genhtml_function_coverage=1 00:05:42.716 --rc genhtml_legend=1 00:05:42.716 --rc geninfo_all_blocks=1 00:05:42.716 --rc geninfo_unexecuted_blocks=1 00:05:42.716 00:05:42.716 ' 00:05:42.716 05:33:15 json_config_extra_key -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:42.716 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.716 --rc genhtml_branch_coverage=1 00:05:42.716 --rc genhtml_function_coverage=1 00:05:42.716 --rc genhtml_legend=1 00:05:42.716 --rc geninfo_all_blocks=1 00:05:42.716 --rc geninfo_unexecuted_blocks=1 00:05:42.716 00:05:42.716 ' 00:05:42.716 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:1b6716e3-b0c8-480d-9903-786aaf87e8fc 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=1b6716e3-b0c8-480d-9903-786aaf87e8fc 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:42.716 05:33:15 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:42.716 05:33:15 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:42.716 05:33:15 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.716 05:33:15 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.716 05:33:15 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.717 05:33:15 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:05:42.717 05:33:15 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.717 05:33:15 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:05:42.717 05:33:15 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:42.717 05:33:15 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:42.717 05:33:15 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:42.717 05:33:15 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:42.717 05:33:15 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:42.717 05:33:15 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:42.717 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:42.717 05:33:15 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:42.717 05:33:15 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:42.717 05:33:15 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:05:42.717 INFO: launching applications... 00:05:42.717 05:33:15 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:05:42.717 05:33:15 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:05:42.717 05:33:15 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:05:42.717 05:33:15 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:05:42.717 05:33:15 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:05:42.717 05:33:15 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:05:42.717 05:33:15 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:05:42.717 05:33:15 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:05:42.717 05:33:15 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=69465 00:05:42.717 05:33:15 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:05:42.717 Waiting for target to run... 00:05:42.717 05:33:15 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 69465 /var/tmp/spdk_tgt.sock 00:05:42.717 05:33:15 json_config_extra_key -- common/autotest_common.sh@835 -- # '[' -z 69465 ']' 00:05:42.717 05:33:15 json_config_extra_key -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:05:42.717 05:33:15 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:05:42.717 05:33:15 json_config_extra_key -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:42.717 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:05:42.717 05:33:15 json_config_extra_key -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:05:42.717 05:33:15 json_config_extra_key -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:42.717 05:33:15 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:05:42.717 [2024-12-07 05:33:16.052012] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:42.717 [2024-12-07 05:33:16.052175] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69465 ] 00:05:43.286 [2024-12-07 05:33:16.415745] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:43.286 [2024-12-07 05:33:16.432795] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:43.545 00:05:43.545 INFO: shutting down applications... 00:05:43.545 05:33:16 json_config_extra_key -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:43.545 05:33:16 json_config_extra_key -- common/autotest_common.sh@868 -- # return 0 00:05:43.545 05:33:16 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:05:43.545 05:33:16 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:05:43.545 05:33:16 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:05:43.545 05:33:16 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:05:43.545 05:33:16 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:05:43.545 05:33:16 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 69465 ]] 00:05:43.545 05:33:16 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 69465 00:05:43.545 05:33:16 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:05:43.545 05:33:16 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:05:43.545 05:33:16 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69465 00:05:43.545 05:33:16 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:05:44.115 05:33:17 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:05:44.115 05:33:17 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:05:44.115 05:33:17 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69465 00:05:44.115 05:33:17 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:05:44.115 05:33:17 json_config_extra_key -- json_config/common.sh@43 -- # break 00:05:44.115 05:33:17 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:05:44.115 05:33:17 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:05:44.115 SPDK target shutdown done 00:05:44.115 05:33:17 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:05:44.115 Success 00:05:44.115 00:05:44.115 real 0m1.650s 00:05:44.115 user 0m1.362s 00:05:44.115 sys 0m0.448s 00:05:44.115 05:33:17 json_config_extra_key -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:44.115 05:33:17 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:05:44.115 ************************************ 00:05:44.115 END TEST json_config_extra_key 00:05:44.115 ************************************ 00:05:44.115 05:33:17 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:05:44.115 05:33:17 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:44.115 05:33:17 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:44.115 05:33:17 -- common/autotest_common.sh@10 -- # set +x 00:05:44.115 ************************************ 00:05:44.115 START TEST alias_rpc 00:05:44.115 ************************************ 00:05:44.115 05:33:17 alias_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:05:44.376 * Looking for test storage... 00:05:44.376 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@1711 -- # lcov --version 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@345 -- # : 1 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:44.376 05:33:17 alias_rpc -- scripts/common.sh@368 -- # return 0 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:44.376 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:44.376 --rc genhtml_branch_coverage=1 00:05:44.376 --rc genhtml_function_coverage=1 00:05:44.376 --rc genhtml_legend=1 00:05:44.376 --rc geninfo_all_blocks=1 00:05:44.376 --rc geninfo_unexecuted_blocks=1 00:05:44.376 00:05:44.376 ' 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:44.376 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:44.376 --rc genhtml_branch_coverage=1 00:05:44.376 --rc genhtml_function_coverage=1 00:05:44.376 --rc genhtml_legend=1 00:05:44.376 --rc geninfo_all_blocks=1 00:05:44.376 --rc geninfo_unexecuted_blocks=1 00:05:44.376 00:05:44.376 ' 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:44.376 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:44.376 --rc genhtml_branch_coverage=1 00:05:44.376 --rc genhtml_function_coverage=1 00:05:44.376 --rc genhtml_legend=1 00:05:44.376 --rc geninfo_all_blocks=1 00:05:44.376 --rc geninfo_unexecuted_blocks=1 00:05:44.376 00:05:44.376 ' 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:44.376 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:44.376 --rc genhtml_branch_coverage=1 00:05:44.376 --rc genhtml_function_coverage=1 00:05:44.376 --rc genhtml_legend=1 00:05:44.376 --rc geninfo_all_blocks=1 00:05:44.376 --rc geninfo_unexecuted_blocks=1 00:05:44.376 00:05:44.376 ' 00:05:44.376 05:33:17 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:05:44.376 05:33:17 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=69544 00:05:44.376 05:33:17 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:44.376 05:33:17 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 69544 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@835 -- # '[' -z 69544 ']' 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:44.376 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:44.376 05:33:17 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:44.636 [2024-12-07 05:33:17.757054] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:44.636 [2024-12-07 05:33:17.757273] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69544 ] 00:05:44.636 [2024-12-07 05:33:17.911708] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:44.636 [2024-12-07 05:33:17.935943] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:45.207 05:33:18 alias_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:45.207 05:33:18 alias_rpc -- common/autotest_common.sh@868 -- # return 0 00:05:45.207 05:33:18 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:05:45.467 05:33:18 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 69544 00:05:45.467 05:33:18 alias_rpc -- common/autotest_common.sh@954 -- # '[' -z 69544 ']' 00:05:45.467 05:33:18 alias_rpc -- common/autotest_common.sh@958 -- # kill -0 69544 00:05:45.467 05:33:18 alias_rpc -- common/autotest_common.sh@959 -- # uname 00:05:45.467 05:33:18 alias_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:45.467 05:33:18 alias_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69544 00:05:45.467 killing process with pid 69544 00:05:45.467 05:33:18 alias_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:45.467 05:33:18 alias_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:45.467 05:33:18 alias_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69544' 00:05:45.467 05:33:18 alias_rpc -- common/autotest_common.sh@973 -- # kill 69544 00:05:45.467 05:33:18 alias_rpc -- common/autotest_common.sh@978 -- # wait 69544 00:05:46.036 ************************************ 00:05:46.036 END TEST alias_rpc 00:05:46.036 ************************************ 00:05:46.036 00:05:46.036 real 0m1.709s 00:05:46.036 user 0m1.721s 00:05:46.036 sys 0m0.469s 00:05:46.036 05:33:19 alias_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:46.036 05:33:19 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:46.036 05:33:19 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:05:46.036 05:33:19 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:05:46.036 05:33:19 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:46.036 05:33:19 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:46.036 05:33:19 -- common/autotest_common.sh@10 -- # set +x 00:05:46.036 ************************************ 00:05:46.036 START TEST spdkcli_tcp 00:05:46.036 ************************************ 00:05:46.036 05:33:19 spdkcli_tcp -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:05:46.036 * Looking for test storage... 00:05:46.036 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:05:46.036 05:33:19 spdkcli_tcp -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:46.036 05:33:19 spdkcli_tcp -- common/autotest_common.sh@1711 -- # lcov --version 00:05:46.036 05:33:19 spdkcli_tcp -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:46.296 05:33:19 spdkcli_tcp -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:46.296 05:33:19 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:05:46.296 05:33:19 spdkcli_tcp -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:46.296 05:33:19 spdkcli_tcp -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:46.296 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.296 --rc genhtml_branch_coverage=1 00:05:46.296 --rc genhtml_function_coverage=1 00:05:46.296 --rc genhtml_legend=1 00:05:46.296 --rc geninfo_all_blocks=1 00:05:46.296 --rc geninfo_unexecuted_blocks=1 00:05:46.296 00:05:46.296 ' 00:05:46.296 05:33:19 spdkcli_tcp -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:46.296 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.296 --rc genhtml_branch_coverage=1 00:05:46.296 --rc genhtml_function_coverage=1 00:05:46.296 --rc genhtml_legend=1 00:05:46.296 --rc geninfo_all_blocks=1 00:05:46.296 --rc geninfo_unexecuted_blocks=1 00:05:46.296 00:05:46.296 ' 00:05:46.296 05:33:19 spdkcli_tcp -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:46.296 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.296 --rc genhtml_branch_coverage=1 00:05:46.296 --rc genhtml_function_coverage=1 00:05:46.296 --rc genhtml_legend=1 00:05:46.296 --rc geninfo_all_blocks=1 00:05:46.296 --rc geninfo_unexecuted_blocks=1 00:05:46.296 00:05:46.296 ' 00:05:46.296 05:33:19 spdkcli_tcp -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:46.296 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.296 --rc genhtml_branch_coverage=1 00:05:46.296 --rc genhtml_function_coverage=1 00:05:46.296 --rc genhtml_legend=1 00:05:46.296 --rc geninfo_all_blocks=1 00:05:46.296 --rc geninfo_unexecuted_blocks=1 00:05:46.296 00:05:46.296 ' 00:05:46.296 05:33:19 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:05:46.296 05:33:19 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:05:46.296 05:33:19 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:05:46.296 05:33:19 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:05:46.296 05:33:19 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:05:46.296 05:33:19 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:05:46.296 05:33:19 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:05:46.296 05:33:19 spdkcli_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:46.296 05:33:19 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:46.296 05:33:19 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=69622 00:05:46.297 05:33:19 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:05:46.297 05:33:19 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 69622 00:05:46.297 05:33:19 spdkcli_tcp -- common/autotest_common.sh@835 -- # '[' -z 69622 ']' 00:05:46.297 05:33:19 spdkcli_tcp -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:46.297 05:33:19 spdkcli_tcp -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:46.297 05:33:19 spdkcli_tcp -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:46.297 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:46.297 05:33:19 spdkcli_tcp -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:46.297 05:33:19 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:46.297 [2024-12-07 05:33:19.536671] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:46.297 [2024-12-07 05:33:19.536845] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69622 ] 00:05:46.556 [2024-12-07 05:33:19.691840] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:46.556 [2024-12-07 05:33:19.718363] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:46.556 [2024-12-07 05:33:19.718473] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:47.147 05:33:20 spdkcli_tcp -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:47.147 05:33:20 spdkcli_tcp -- common/autotest_common.sh@868 -- # return 0 00:05:47.147 05:33:20 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=69635 00:05:47.147 05:33:20 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:05:47.147 05:33:20 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:05:47.408 [ 00:05:47.408 "bdev_malloc_delete", 00:05:47.408 "bdev_malloc_create", 00:05:47.408 "bdev_null_resize", 00:05:47.408 "bdev_null_delete", 00:05:47.408 "bdev_null_create", 00:05:47.408 "bdev_nvme_cuse_unregister", 00:05:47.408 "bdev_nvme_cuse_register", 00:05:47.408 "bdev_opal_new_user", 00:05:47.408 "bdev_opal_set_lock_state", 00:05:47.408 "bdev_opal_delete", 00:05:47.408 "bdev_opal_get_info", 00:05:47.408 "bdev_opal_create", 00:05:47.408 "bdev_nvme_opal_revert", 00:05:47.408 "bdev_nvme_opal_init", 00:05:47.408 "bdev_nvme_send_cmd", 00:05:47.408 "bdev_nvme_set_keys", 00:05:47.408 "bdev_nvme_get_path_iostat", 00:05:47.408 "bdev_nvme_get_mdns_discovery_info", 00:05:47.408 "bdev_nvme_stop_mdns_discovery", 00:05:47.408 "bdev_nvme_start_mdns_discovery", 00:05:47.408 "bdev_nvme_set_multipath_policy", 00:05:47.408 "bdev_nvme_set_preferred_path", 00:05:47.408 "bdev_nvme_get_io_paths", 00:05:47.408 "bdev_nvme_remove_error_injection", 00:05:47.408 "bdev_nvme_add_error_injection", 00:05:47.408 "bdev_nvme_get_discovery_info", 00:05:47.408 "bdev_nvme_stop_discovery", 00:05:47.408 "bdev_nvme_start_discovery", 00:05:47.408 "bdev_nvme_get_controller_health_info", 00:05:47.408 "bdev_nvme_disable_controller", 00:05:47.408 "bdev_nvme_enable_controller", 00:05:47.408 "bdev_nvme_reset_controller", 00:05:47.408 "bdev_nvme_get_transport_statistics", 00:05:47.408 "bdev_nvme_apply_firmware", 00:05:47.408 "bdev_nvme_detach_controller", 00:05:47.408 "bdev_nvme_get_controllers", 00:05:47.408 "bdev_nvme_attach_controller", 00:05:47.408 "bdev_nvme_set_hotplug", 00:05:47.408 "bdev_nvme_set_options", 00:05:47.408 "bdev_passthru_delete", 00:05:47.408 "bdev_passthru_create", 00:05:47.408 "bdev_lvol_set_parent_bdev", 00:05:47.408 "bdev_lvol_set_parent", 00:05:47.408 "bdev_lvol_check_shallow_copy", 00:05:47.408 "bdev_lvol_start_shallow_copy", 00:05:47.408 "bdev_lvol_grow_lvstore", 00:05:47.408 "bdev_lvol_get_lvols", 00:05:47.408 "bdev_lvol_get_lvstores", 00:05:47.408 "bdev_lvol_delete", 00:05:47.408 "bdev_lvol_set_read_only", 00:05:47.408 "bdev_lvol_resize", 00:05:47.408 "bdev_lvol_decouple_parent", 00:05:47.408 "bdev_lvol_inflate", 00:05:47.408 "bdev_lvol_rename", 00:05:47.408 "bdev_lvol_clone_bdev", 00:05:47.408 "bdev_lvol_clone", 00:05:47.408 "bdev_lvol_snapshot", 00:05:47.408 "bdev_lvol_create", 00:05:47.408 "bdev_lvol_delete_lvstore", 00:05:47.408 "bdev_lvol_rename_lvstore", 00:05:47.408 "bdev_lvol_create_lvstore", 00:05:47.408 "bdev_raid_set_options", 00:05:47.408 "bdev_raid_remove_base_bdev", 00:05:47.408 "bdev_raid_add_base_bdev", 00:05:47.408 "bdev_raid_delete", 00:05:47.408 "bdev_raid_create", 00:05:47.408 "bdev_raid_get_bdevs", 00:05:47.408 "bdev_error_inject_error", 00:05:47.408 "bdev_error_delete", 00:05:47.408 "bdev_error_create", 00:05:47.408 "bdev_split_delete", 00:05:47.408 "bdev_split_create", 00:05:47.408 "bdev_delay_delete", 00:05:47.408 "bdev_delay_create", 00:05:47.408 "bdev_delay_update_latency", 00:05:47.408 "bdev_zone_block_delete", 00:05:47.408 "bdev_zone_block_create", 00:05:47.408 "blobfs_create", 00:05:47.408 "blobfs_detect", 00:05:47.408 "blobfs_set_cache_size", 00:05:47.408 "bdev_aio_delete", 00:05:47.408 "bdev_aio_rescan", 00:05:47.408 "bdev_aio_create", 00:05:47.408 "bdev_ftl_set_property", 00:05:47.408 "bdev_ftl_get_properties", 00:05:47.408 "bdev_ftl_get_stats", 00:05:47.408 "bdev_ftl_unmap", 00:05:47.408 "bdev_ftl_unload", 00:05:47.408 "bdev_ftl_delete", 00:05:47.408 "bdev_ftl_load", 00:05:47.408 "bdev_ftl_create", 00:05:47.408 "bdev_virtio_attach_controller", 00:05:47.408 "bdev_virtio_scsi_get_devices", 00:05:47.408 "bdev_virtio_detach_controller", 00:05:47.408 "bdev_virtio_blk_set_hotplug", 00:05:47.408 "bdev_iscsi_delete", 00:05:47.408 "bdev_iscsi_create", 00:05:47.408 "bdev_iscsi_set_options", 00:05:47.408 "accel_error_inject_error", 00:05:47.408 "ioat_scan_accel_module", 00:05:47.408 "dsa_scan_accel_module", 00:05:47.408 "iaa_scan_accel_module", 00:05:47.408 "keyring_file_remove_key", 00:05:47.408 "keyring_file_add_key", 00:05:47.408 "keyring_linux_set_options", 00:05:47.408 "fsdev_aio_delete", 00:05:47.408 "fsdev_aio_create", 00:05:47.408 "iscsi_get_histogram", 00:05:47.408 "iscsi_enable_histogram", 00:05:47.408 "iscsi_set_options", 00:05:47.408 "iscsi_get_auth_groups", 00:05:47.408 "iscsi_auth_group_remove_secret", 00:05:47.408 "iscsi_auth_group_add_secret", 00:05:47.408 "iscsi_delete_auth_group", 00:05:47.408 "iscsi_create_auth_group", 00:05:47.408 "iscsi_set_discovery_auth", 00:05:47.408 "iscsi_get_options", 00:05:47.408 "iscsi_target_node_request_logout", 00:05:47.408 "iscsi_target_node_set_redirect", 00:05:47.408 "iscsi_target_node_set_auth", 00:05:47.408 "iscsi_target_node_add_lun", 00:05:47.409 "iscsi_get_stats", 00:05:47.409 "iscsi_get_connections", 00:05:47.409 "iscsi_portal_group_set_auth", 00:05:47.409 "iscsi_start_portal_group", 00:05:47.409 "iscsi_delete_portal_group", 00:05:47.409 "iscsi_create_portal_group", 00:05:47.409 "iscsi_get_portal_groups", 00:05:47.409 "iscsi_delete_target_node", 00:05:47.409 "iscsi_target_node_remove_pg_ig_maps", 00:05:47.409 "iscsi_target_node_add_pg_ig_maps", 00:05:47.409 "iscsi_create_target_node", 00:05:47.409 "iscsi_get_target_nodes", 00:05:47.409 "iscsi_delete_initiator_group", 00:05:47.409 "iscsi_initiator_group_remove_initiators", 00:05:47.409 "iscsi_initiator_group_add_initiators", 00:05:47.409 "iscsi_create_initiator_group", 00:05:47.409 "iscsi_get_initiator_groups", 00:05:47.409 "nvmf_set_crdt", 00:05:47.409 "nvmf_set_config", 00:05:47.409 "nvmf_set_max_subsystems", 00:05:47.409 "nvmf_stop_mdns_prr", 00:05:47.409 "nvmf_publish_mdns_prr", 00:05:47.409 "nvmf_subsystem_get_listeners", 00:05:47.409 "nvmf_subsystem_get_qpairs", 00:05:47.409 "nvmf_subsystem_get_controllers", 00:05:47.409 "nvmf_get_stats", 00:05:47.409 "nvmf_get_transports", 00:05:47.409 "nvmf_create_transport", 00:05:47.409 "nvmf_get_targets", 00:05:47.409 "nvmf_delete_target", 00:05:47.409 "nvmf_create_target", 00:05:47.409 "nvmf_subsystem_allow_any_host", 00:05:47.409 "nvmf_subsystem_set_keys", 00:05:47.409 "nvmf_subsystem_remove_host", 00:05:47.409 "nvmf_subsystem_add_host", 00:05:47.409 "nvmf_ns_remove_host", 00:05:47.409 "nvmf_ns_add_host", 00:05:47.409 "nvmf_subsystem_remove_ns", 00:05:47.409 "nvmf_subsystem_set_ns_ana_group", 00:05:47.409 "nvmf_subsystem_add_ns", 00:05:47.409 "nvmf_subsystem_listener_set_ana_state", 00:05:47.409 "nvmf_discovery_get_referrals", 00:05:47.409 "nvmf_discovery_remove_referral", 00:05:47.409 "nvmf_discovery_add_referral", 00:05:47.409 "nvmf_subsystem_remove_listener", 00:05:47.409 "nvmf_subsystem_add_listener", 00:05:47.409 "nvmf_delete_subsystem", 00:05:47.409 "nvmf_create_subsystem", 00:05:47.409 "nvmf_get_subsystems", 00:05:47.409 "env_dpdk_get_mem_stats", 00:05:47.409 "nbd_get_disks", 00:05:47.409 "nbd_stop_disk", 00:05:47.409 "nbd_start_disk", 00:05:47.409 "ublk_recover_disk", 00:05:47.409 "ublk_get_disks", 00:05:47.409 "ublk_stop_disk", 00:05:47.409 "ublk_start_disk", 00:05:47.409 "ublk_destroy_target", 00:05:47.409 "ublk_create_target", 00:05:47.409 "virtio_blk_create_transport", 00:05:47.409 "virtio_blk_get_transports", 00:05:47.409 "vhost_controller_set_coalescing", 00:05:47.409 "vhost_get_controllers", 00:05:47.409 "vhost_delete_controller", 00:05:47.409 "vhost_create_blk_controller", 00:05:47.409 "vhost_scsi_controller_remove_target", 00:05:47.409 "vhost_scsi_controller_add_target", 00:05:47.409 "vhost_start_scsi_controller", 00:05:47.409 "vhost_create_scsi_controller", 00:05:47.409 "thread_set_cpumask", 00:05:47.409 "scheduler_set_options", 00:05:47.409 "framework_get_governor", 00:05:47.409 "framework_get_scheduler", 00:05:47.409 "framework_set_scheduler", 00:05:47.409 "framework_get_reactors", 00:05:47.409 "thread_get_io_channels", 00:05:47.409 "thread_get_pollers", 00:05:47.409 "thread_get_stats", 00:05:47.409 "framework_monitor_context_switch", 00:05:47.409 "spdk_kill_instance", 00:05:47.409 "log_enable_timestamps", 00:05:47.409 "log_get_flags", 00:05:47.409 "log_clear_flag", 00:05:47.409 "log_set_flag", 00:05:47.409 "log_get_level", 00:05:47.409 "log_set_level", 00:05:47.409 "log_get_print_level", 00:05:47.409 "log_set_print_level", 00:05:47.409 "framework_enable_cpumask_locks", 00:05:47.409 "framework_disable_cpumask_locks", 00:05:47.409 "framework_wait_init", 00:05:47.409 "framework_start_init", 00:05:47.409 "scsi_get_devices", 00:05:47.409 "bdev_get_histogram", 00:05:47.409 "bdev_enable_histogram", 00:05:47.409 "bdev_set_qos_limit", 00:05:47.409 "bdev_set_qd_sampling_period", 00:05:47.409 "bdev_get_bdevs", 00:05:47.409 "bdev_reset_iostat", 00:05:47.409 "bdev_get_iostat", 00:05:47.409 "bdev_examine", 00:05:47.409 "bdev_wait_for_examine", 00:05:47.409 "bdev_set_options", 00:05:47.409 "accel_get_stats", 00:05:47.409 "accel_set_options", 00:05:47.409 "accel_set_driver", 00:05:47.409 "accel_crypto_key_destroy", 00:05:47.409 "accel_crypto_keys_get", 00:05:47.409 "accel_crypto_key_create", 00:05:47.409 "accel_assign_opc", 00:05:47.409 "accel_get_module_info", 00:05:47.409 "accel_get_opc_assignments", 00:05:47.409 "vmd_rescan", 00:05:47.409 "vmd_remove_device", 00:05:47.409 "vmd_enable", 00:05:47.409 "sock_get_default_impl", 00:05:47.409 "sock_set_default_impl", 00:05:47.409 "sock_impl_set_options", 00:05:47.409 "sock_impl_get_options", 00:05:47.409 "iobuf_get_stats", 00:05:47.409 "iobuf_set_options", 00:05:47.409 "keyring_get_keys", 00:05:47.409 "framework_get_pci_devices", 00:05:47.409 "framework_get_config", 00:05:47.409 "framework_get_subsystems", 00:05:47.409 "fsdev_set_opts", 00:05:47.409 "fsdev_get_opts", 00:05:47.409 "trace_get_info", 00:05:47.409 "trace_get_tpoint_group_mask", 00:05:47.409 "trace_disable_tpoint_group", 00:05:47.409 "trace_enable_tpoint_group", 00:05:47.409 "trace_clear_tpoint_mask", 00:05:47.409 "trace_set_tpoint_mask", 00:05:47.409 "notify_get_notifications", 00:05:47.409 "notify_get_types", 00:05:47.409 "spdk_get_version", 00:05:47.409 "rpc_get_methods" 00:05:47.409 ] 00:05:47.409 05:33:20 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:47.409 05:33:20 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:05:47.409 05:33:20 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 69622 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@954 -- # '[' -z 69622 ']' 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@958 -- # kill -0 69622 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@959 -- # uname 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69622 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69622' 00:05:47.409 killing process with pid 69622 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@973 -- # kill 69622 00:05:47.409 05:33:20 spdkcli_tcp -- common/autotest_common.sh@978 -- # wait 69622 00:05:47.669 ************************************ 00:05:47.669 END TEST spdkcli_tcp 00:05:47.669 ************************************ 00:05:47.669 00:05:47.669 real 0m1.796s 00:05:47.669 user 0m2.993s 00:05:47.669 sys 0m0.583s 00:05:47.669 05:33:21 spdkcli_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:47.669 05:33:21 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:47.928 05:33:21 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:05:47.928 05:33:21 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:47.928 05:33:21 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:47.928 05:33:21 -- common/autotest_common.sh@10 -- # set +x 00:05:47.928 ************************************ 00:05:47.928 START TEST dpdk_mem_utility 00:05:47.928 ************************************ 00:05:47.928 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:05:47.928 * Looking for test storage... 00:05:47.928 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:05:47.928 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:47.928 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@1711 -- # lcov --version 00:05:47.928 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:47.928 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:05:47.928 05:33:21 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:05:47.929 05:33:21 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:47.929 05:33:21 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:05:47.929 05:33:21 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:05:47.929 05:33:21 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:47.929 05:33:21 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:47.929 05:33:21 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:05:47.929 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:47.929 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:47.929 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:47.929 --rc genhtml_branch_coverage=1 00:05:47.929 --rc genhtml_function_coverage=1 00:05:47.929 --rc genhtml_legend=1 00:05:47.929 --rc geninfo_all_blocks=1 00:05:47.929 --rc geninfo_unexecuted_blocks=1 00:05:47.929 00:05:47.929 ' 00:05:48.187 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:48.187 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.187 --rc genhtml_branch_coverage=1 00:05:48.187 --rc genhtml_function_coverage=1 00:05:48.187 --rc genhtml_legend=1 00:05:48.187 --rc geninfo_all_blocks=1 00:05:48.187 --rc geninfo_unexecuted_blocks=1 00:05:48.187 00:05:48.187 ' 00:05:48.187 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:48.187 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.187 --rc genhtml_branch_coverage=1 00:05:48.187 --rc genhtml_function_coverage=1 00:05:48.187 --rc genhtml_legend=1 00:05:48.187 --rc geninfo_all_blocks=1 00:05:48.187 --rc geninfo_unexecuted_blocks=1 00:05:48.187 00:05:48.187 ' 00:05:48.187 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:48.187 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.188 --rc genhtml_branch_coverage=1 00:05:48.188 --rc genhtml_function_coverage=1 00:05:48.188 --rc genhtml_legend=1 00:05:48.188 --rc geninfo_all_blocks=1 00:05:48.188 --rc geninfo_unexecuted_blocks=1 00:05:48.188 00:05:48.188 ' 00:05:48.188 05:33:21 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:05:48.188 05:33:21 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=69718 00:05:48.188 05:33:21 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:48.188 05:33:21 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 69718 00:05:48.188 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@835 -- # '[' -z 69718 ']' 00:05:48.188 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:48.188 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:48.188 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:48.188 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:48.188 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:48.188 05:33:21 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:48.188 [2024-12-07 05:33:21.388302] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:48.188 [2024-12-07 05:33:21.388874] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69718 ] 00:05:48.188 [2024-12-07 05:33:21.543243] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:48.446 [2024-12-07 05:33:21.570017] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:49.015 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:49.015 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@868 -- # return 0 00:05:49.015 05:33:22 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:05:49.015 05:33:22 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:05:49.015 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:49.015 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:49.015 { 00:05:49.015 "filename": "/tmp/spdk_mem_dump.txt" 00:05:49.015 } 00:05:49.015 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:49.015 05:33:22 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:05:49.015 DPDK memory size 818.000000 MiB in 1 heap(s) 00:05:49.015 1 heaps totaling size 818.000000 MiB 00:05:49.015 size: 818.000000 MiB heap id: 0 00:05:49.015 end heaps---------- 00:05:49.015 9 mempools totaling size 603.782043 MiB 00:05:49.015 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:05:49.015 size: 158.602051 MiB name: PDU_data_out_Pool 00:05:49.015 size: 100.555481 MiB name: bdev_io_69718 00:05:49.015 size: 50.003479 MiB name: msgpool_69718 00:05:49.015 size: 36.509338 MiB name: fsdev_io_69718 00:05:49.015 size: 21.763794 MiB name: PDU_Pool 00:05:49.015 size: 19.513306 MiB name: SCSI_TASK_Pool 00:05:49.015 size: 4.133484 MiB name: evtpool_69718 00:05:49.015 size: 0.026123 MiB name: Session_Pool 00:05:49.015 end mempools------- 00:05:49.015 6 memzones totaling size 4.142822 MiB 00:05:49.015 size: 1.000366 MiB name: RG_ring_0_69718 00:05:49.015 size: 1.000366 MiB name: RG_ring_1_69718 00:05:49.015 size: 1.000366 MiB name: RG_ring_4_69718 00:05:49.015 size: 1.000366 MiB name: RG_ring_5_69718 00:05:49.015 size: 0.125366 MiB name: RG_ring_2_69718 00:05:49.015 size: 0.015991 MiB name: RG_ring_3_69718 00:05:49.015 end memzones------- 00:05:49.015 05:33:22 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:05:49.015 heap id: 0 total size: 818.000000 MiB number of busy elements: 309 number of free elements: 15 00:05:49.015 list of free elements. size: 10.803955 MiB 00:05:49.015 element at address: 0x200019200000 with size: 0.999878 MiB 00:05:49.015 element at address: 0x200019400000 with size: 0.999878 MiB 00:05:49.015 element at address: 0x200032000000 with size: 0.994446 MiB 00:05:49.015 element at address: 0x200000400000 with size: 0.993958 MiB 00:05:49.015 element at address: 0x200006400000 with size: 0.959839 MiB 00:05:49.015 element at address: 0x200012c00000 with size: 0.944275 MiB 00:05:49.015 element at address: 0x200019600000 with size: 0.936584 MiB 00:05:49.015 element at address: 0x200000200000 with size: 0.717346 MiB 00:05:49.015 element at address: 0x20001ae00000 with size: 0.568970 MiB 00:05:49.015 element at address: 0x20000a600000 with size: 0.488892 MiB 00:05:49.015 element at address: 0x200000c00000 with size: 0.486267 MiB 00:05:49.015 element at address: 0x200019800000 with size: 0.485657 MiB 00:05:49.015 element at address: 0x200003e00000 with size: 0.480286 MiB 00:05:49.015 element at address: 0x200028200000 with size: 0.395935 MiB 00:05:49.015 element at address: 0x200000800000 with size: 0.351746 MiB 00:05:49.015 list of standard malloc elements. size: 199.267151 MiB 00:05:49.015 element at address: 0x20000a7fff80 with size: 132.000122 MiB 00:05:49.015 element at address: 0x2000065fff80 with size: 64.000122 MiB 00:05:49.015 element at address: 0x2000192fff80 with size: 1.000122 MiB 00:05:49.015 element at address: 0x2000194fff80 with size: 1.000122 MiB 00:05:49.015 element at address: 0x2000196fff80 with size: 1.000122 MiB 00:05:49.016 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:05:49.016 element at address: 0x2000196eff00 with size: 0.062622 MiB 00:05:49.016 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:05:49.016 element at address: 0x2000196efdc0 with size: 0.000305 MiB 00:05:49.016 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004fe740 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004fe800 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004fe8c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004fe980 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004fea40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004feb00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004febc0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004fec80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004fed40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004fee00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004feec0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004fef80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff040 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff100 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff1c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff280 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff340 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff400 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff4c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff580 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff640 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff700 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff7c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff880 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ff940 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ffa00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ffac0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ffcc0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ffd80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000004ffe40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000085a0c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000085a2c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000085e580 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087e840 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087e900 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087e9c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087ea80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087eb40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087ec00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087ecc0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087ed80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087ee40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087ef00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087efc0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087f080 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087f140 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087f200 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087f2c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087f380 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087f440 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087f500 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087f5c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000087f680 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000008ff940 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000008ffb40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7c7c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7c880 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7c940 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7ca00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7cac0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7cb80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7cc40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7cd00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7cdc0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7ce80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7cf40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d000 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d0c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d180 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d240 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d300 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d3c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d480 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d540 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d600 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d6c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d780 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d840 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d900 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7d9c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7da80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7db40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7dc00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7dcc0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7dd80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7de40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7df00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7dfc0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e080 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e140 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e200 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e2c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e380 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e440 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e500 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e5c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e680 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e740 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e800 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e8c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7e980 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7ea40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7eb00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7ebc0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7ec80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000c7ed40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000cff000 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200000cff0c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003e7af40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003e7b000 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003e7b0c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003e7b180 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003e7b240 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003e7b300 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003e7b3c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003e7b480 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003e7b540 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003e7b600 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003e7b6c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200003efb980 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000064fdd80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67d280 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67d340 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67d400 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67d4c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67d580 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67d640 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67d700 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67d7c0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67d880 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67d940 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67da00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a67dac0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20000a6fdd80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x200012cf1bc0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000196efc40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000196efd00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x2000198bc740 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20001ae91a80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20001ae91b40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20001ae91c00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20001ae91cc0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20001ae91d80 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20001ae91e40 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20001ae91f00 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20001ae91fc0 with size: 0.000183 MiB 00:05:49.016 element at address: 0x20001ae92080 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92140 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92200 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae922c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92380 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92440 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92500 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae925c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92680 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92740 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92800 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae928c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92980 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92a40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92b00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92bc0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92c80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92d40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92e00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92ec0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae92f80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93040 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93100 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae931c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93280 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93340 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93400 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae934c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93580 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93640 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93700 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae937c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93880 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93940 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93a00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93ac0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93b80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93c40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93d00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93dc0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93e80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae93f40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94000 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae940c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94180 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94240 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94300 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae943c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94480 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94540 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94600 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae946c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94780 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94840 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94900 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae949c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94a80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94b40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94c00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94cc0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94d80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94e40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94f00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae94fc0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae95080 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae95140 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae95200 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae952c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae95380 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20001ae95440 with size: 0.000183 MiB 00:05:49.017 element at address: 0x2000282655c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x200028265680 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826c280 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826c480 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826c540 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826c600 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826c6c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826c780 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826c840 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826c900 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826c9c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826ca80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826cb40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826cc00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826ccc0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826cd80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826ce40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826cf00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826cfc0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d080 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d140 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d200 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d2c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d380 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d440 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d500 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d5c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d680 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d740 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d800 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d8c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826d980 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826da40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826db00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826dbc0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826dc80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826dd40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826de00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826dec0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826df80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e040 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e100 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e1c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e280 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e340 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e400 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e4c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e580 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e640 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e700 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e7c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e880 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826e940 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826ea00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826eac0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826eb80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826ec40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826ed00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826edc0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826ee80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826ef40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f000 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f0c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f180 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f240 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f300 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f3c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f480 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f540 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f600 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f6c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f780 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f840 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f900 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826f9c0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826fa80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826fb40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826fc00 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826fcc0 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826fd80 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826fe40 with size: 0.000183 MiB 00:05:49.017 element at address: 0x20002826ff00 with size: 0.000183 MiB 00:05:49.017 list of memzone associated elements. size: 607.928894 MiB 00:05:49.017 element at address: 0x20001ae95500 with size: 211.416748 MiB 00:05:49.017 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:05:49.018 element at address: 0x20002826ffc0 with size: 157.562561 MiB 00:05:49.018 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:05:49.018 element at address: 0x200012df1e80 with size: 100.055054 MiB 00:05:49.018 associated memzone info: size: 100.054932 MiB name: MP_bdev_io_69718_0 00:05:49.018 element at address: 0x200000dff380 with size: 48.003052 MiB 00:05:49.018 associated memzone info: size: 48.002930 MiB name: MP_msgpool_69718_0 00:05:49.018 element at address: 0x200003ffdb80 with size: 36.008911 MiB 00:05:49.018 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_69718_0 00:05:49.018 element at address: 0x2000199be940 with size: 20.255554 MiB 00:05:49.018 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:05:49.018 element at address: 0x2000321feb40 with size: 18.005066 MiB 00:05:49.018 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:05:49.018 element at address: 0x2000004fff00 with size: 3.000244 MiB 00:05:49.018 associated memzone info: size: 3.000122 MiB name: MP_evtpool_69718_0 00:05:49.018 element at address: 0x2000009ffe00 with size: 2.000488 MiB 00:05:49.018 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_69718 00:05:49.018 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:05:49.018 associated memzone info: size: 1.007996 MiB name: MP_evtpool_69718 00:05:49.018 element at address: 0x20000a6fde40 with size: 1.008118 MiB 00:05:49.018 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:05:49.018 element at address: 0x2000198bc800 with size: 1.008118 MiB 00:05:49.018 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:05:49.018 element at address: 0x2000064fde40 with size: 1.008118 MiB 00:05:49.018 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:05:49.018 element at address: 0x200003efba40 with size: 1.008118 MiB 00:05:49.018 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:05:49.018 element at address: 0x200000cff180 with size: 1.000488 MiB 00:05:49.018 associated memzone info: size: 1.000366 MiB name: RG_ring_0_69718 00:05:49.018 element at address: 0x2000008ffc00 with size: 1.000488 MiB 00:05:49.018 associated memzone info: size: 1.000366 MiB name: RG_ring_1_69718 00:05:49.018 element at address: 0x200012cf1c80 with size: 1.000488 MiB 00:05:49.018 associated memzone info: size: 1.000366 MiB name: RG_ring_4_69718 00:05:49.018 element at address: 0x2000320fe940 with size: 1.000488 MiB 00:05:49.018 associated memzone info: size: 1.000366 MiB name: RG_ring_5_69718 00:05:49.018 element at address: 0x20000087f740 with size: 0.500488 MiB 00:05:49.018 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_69718 00:05:49.018 element at address: 0x200000c7ee00 with size: 0.500488 MiB 00:05:49.018 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_69718 00:05:49.018 element at address: 0x20000a67db80 with size: 0.500488 MiB 00:05:49.018 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:05:49.018 element at address: 0x200003e7b780 with size: 0.500488 MiB 00:05:49.018 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:05:49.018 element at address: 0x20001987c540 with size: 0.250488 MiB 00:05:49.018 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:05:49.018 element at address: 0x2000002b7a40 with size: 0.125488 MiB 00:05:49.018 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_69718 00:05:49.018 element at address: 0x20000085e640 with size: 0.125488 MiB 00:05:49.018 associated memzone info: size: 0.125366 MiB name: RG_ring_2_69718 00:05:49.018 element at address: 0x2000064f5b80 with size: 0.031738 MiB 00:05:49.018 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:05:49.018 element at address: 0x200028265740 with size: 0.023743 MiB 00:05:49.018 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:05:49.018 element at address: 0x20000085a380 with size: 0.016113 MiB 00:05:49.018 associated memzone info: size: 0.015991 MiB name: RG_ring_3_69718 00:05:49.018 element at address: 0x20002826b880 with size: 0.002441 MiB 00:05:49.018 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:05:49.018 element at address: 0x2000004ffb80 with size: 0.000305 MiB 00:05:49.018 associated memzone info: size: 0.000183 MiB name: MP_msgpool_69718 00:05:49.018 element at address: 0x2000008ffa00 with size: 0.000305 MiB 00:05:49.018 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_69718 00:05:49.018 element at address: 0x20000085a180 with size: 0.000305 MiB 00:05:49.018 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_69718 00:05:49.018 element at address: 0x20002826c340 with size: 0.000305 MiB 00:05:49.018 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:05:49.018 05:33:22 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:05:49.018 05:33:22 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 69718 00:05:49.018 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@954 -- # '[' -z 69718 ']' 00:05:49.018 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@958 -- # kill -0 69718 00:05:49.018 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@959 -- # uname 00:05:49.018 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:49.018 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69718 00:05:49.018 killing process with pid 69718 00:05:49.018 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:49.018 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:49.018 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69718' 00:05:49.018 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@973 -- # kill 69718 00:05:49.018 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@978 -- # wait 69718 00:05:49.586 ************************************ 00:05:49.586 END TEST dpdk_mem_utility 00:05:49.586 ************************************ 00:05:49.586 00:05:49.586 real 0m1.655s 00:05:49.586 user 0m1.611s 00:05:49.586 sys 0m0.486s 00:05:49.586 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:49.586 05:33:22 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:49.586 05:33:22 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:05:49.586 05:33:22 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:49.586 05:33:22 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:49.586 05:33:22 -- common/autotest_common.sh@10 -- # set +x 00:05:49.586 ************************************ 00:05:49.586 START TEST event 00:05:49.586 ************************************ 00:05:49.586 05:33:22 event -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:05:49.586 * Looking for test storage... 00:05:49.586 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:05:49.586 05:33:22 event -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:49.586 05:33:22 event -- common/autotest_common.sh@1711 -- # lcov --version 00:05:49.586 05:33:22 event -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:49.845 05:33:22 event -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:49.846 05:33:22 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:49.846 05:33:22 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:49.846 05:33:22 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:49.846 05:33:22 event -- scripts/common.sh@336 -- # IFS=.-: 00:05:49.846 05:33:22 event -- scripts/common.sh@336 -- # read -ra ver1 00:05:49.846 05:33:22 event -- scripts/common.sh@337 -- # IFS=.-: 00:05:49.846 05:33:22 event -- scripts/common.sh@337 -- # read -ra ver2 00:05:49.846 05:33:22 event -- scripts/common.sh@338 -- # local 'op=<' 00:05:49.846 05:33:22 event -- scripts/common.sh@340 -- # ver1_l=2 00:05:49.846 05:33:22 event -- scripts/common.sh@341 -- # ver2_l=1 00:05:49.846 05:33:22 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:49.846 05:33:22 event -- scripts/common.sh@344 -- # case "$op" in 00:05:49.846 05:33:22 event -- scripts/common.sh@345 -- # : 1 00:05:49.846 05:33:22 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:49.846 05:33:22 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:49.846 05:33:22 event -- scripts/common.sh@365 -- # decimal 1 00:05:49.846 05:33:22 event -- scripts/common.sh@353 -- # local d=1 00:05:49.846 05:33:22 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:49.846 05:33:23 event -- scripts/common.sh@355 -- # echo 1 00:05:49.846 05:33:23 event -- scripts/common.sh@365 -- # ver1[v]=1 00:05:49.846 05:33:23 event -- scripts/common.sh@366 -- # decimal 2 00:05:49.846 05:33:23 event -- scripts/common.sh@353 -- # local d=2 00:05:49.846 05:33:23 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:49.846 05:33:23 event -- scripts/common.sh@355 -- # echo 2 00:05:49.846 05:33:23 event -- scripts/common.sh@366 -- # ver2[v]=2 00:05:49.846 05:33:23 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:49.846 05:33:23 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:49.846 05:33:23 event -- scripts/common.sh@368 -- # return 0 00:05:49.846 05:33:23 event -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:49.846 05:33:23 event -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:49.846 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.846 --rc genhtml_branch_coverage=1 00:05:49.846 --rc genhtml_function_coverage=1 00:05:49.846 --rc genhtml_legend=1 00:05:49.846 --rc geninfo_all_blocks=1 00:05:49.846 --rc geninfo_unexecuted_blocks=1 00:05:49.846 00:05:49.846 ' 00:05:49.846 05:33:23 event -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:49.846 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.846 --rc genhtml_branch_coverage=1 00:05:49.846 --rc genhtml_function_coverage=1 00:05:49.846 --rc genhtml_legend=1 00:05:49.846 --rc geninfo_all_blocks=1 00:05:49.846 --rc geninfo_unexecuted_blocks=1 00:05:49.846 00:05:49.846 ' 00:05:49.846 05:33:23 event -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:49.846 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.846 --rc genhtml_branch_coverage=1 00:05:49.846 --rc genhtml_function_coverage=1 00:05:49.846 --rc genhtml_legend=1 00:05:49.846 --rc geninfo_all_blocks=1 00:05:49.846 --rc geninfo_unexecuted_blocks=1 00:05:49.846 00:05:49.846 ' 00:05:49.846 05:33:23 event -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:49.846 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.846 --rc genhtml_branch_coverage=1 00:05:49.846 --rc genhtml_function_coverage=1 00:05:49.846 --rc genhtml_legend=1 00:05:49.846 --rc geninfo_all_blocks=1 00:05:49.846 --rc geninfo_unexecuted_blocks=1 00:05:49.846 00:05:49.846 ' 00:05:49.846 05:33:23 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:05:49.846 05:33:23 event -- bdev/nbd_common.sh@6 -- # set -e 00:05:49.846 05:33:23 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:05:49.846 05:33:23 event -- common/autotest_common.sh@1105 -- # '[' 6 -le 1 ']' 00:05:49.846 05:33:23 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:49.846 05:33:23 event -- common/autotest_common.sh@10 -- # set +x 00:05:49.846 ************************************ 00:05:49.846 START TEST event_perf 00:05:49.846 ************************************ 00:05:49.846 05:33:23 event.event_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:05:49.846 Running I/O for 1 seconds...[2024-12-07 05:33:23.065557] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:49.846 [2024-12-07 05:33:23.065730] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69804 ] 00:05:50.106 [2024-12-07 05:33:23.220441] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:05:50.106 [2024-12-07 05:33:23.248277] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:50.106 [2024-12-07 05:33:23.248488] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:05:50.106 [2024-12-07 05:33:23.248695] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:05:50.106 [2024-12-07 05:33:23.248511] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:51.047 Running I/O for 1 seconds... 00:05:51.047 lcore 0: 164614 00:05:51.047 lcore 1: 164613 00:05:51.047 lcore 2: 164615 00:05:51.047 lcore 3: 164614 00:05:51.047 done. 00:05:51.047 ************************************ 00:05:51.047 END TEST event_perf 00:05:51.047 ************************************ 00:05:51.047 00:05:51.047 real 0m1.275s 00:05:51.047 user 0m4.084s 00:05:51.047 sys 0m0.072s 00:05:51.047 05:33:24 event.event_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:51.047 05:33:24 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:05:51.047 05:33:24 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:05:51.047 05:33:24 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:05:51.047 05:33:24 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:51.047 05:33:24 event -- common/autotest_common.sh@10 -- # set +x 00:05:51.047 ************************************ 00:05:51.047 START TEST event_reactor 00:05:51.047 ************************************ 00:05:51.047 05:33:24 event.event_reactor -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:05:51.047 [2024-12-07 05:33:24.404103] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:51.047 [2024-12-07 05:33:24.404290] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69838 ] 00:05:51.307 [2024-12-07 05:33:24.556357] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:51.307 [2024-12-07 05:33:24.581363] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:52.686 test_start 00:05:52.686 oneshot 00:05:52.686 tick 100 00:05:52.686 tick 100 00:05:52.686 tick 250 00:05:52.686 tick 100 00:05:52.686 tick 100 00:05:52.686 tick 100 00:05:52.686 tick 250 00:05:52.686 tick 500 00:05:52.686 tick 100 00:05:52.686 tick 100 00:05:52.686 tick 250 00:05:52.686 tick 100 00:05:52.686 tick 100 00:05:52.686 test_end 00:05:52.686 00:05:52.686 real 0m1.275s 00:05:52.686 user 0m1.099s 00:05:52.686 sys 0m0.069s 00:05:52.686 05:33:25 event.event_reactor -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:52.686 05:33:25 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:05:52.686 ************************************ 00:05:52.686 END TEST event_reactor 00:05:52.686 ************************************ 00:05:52.686 05:33:25 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:05:52.686 05:33:25 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:05:52.686 05:33:25 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:52.686 05:33:25 event -- common/autotest_common.sh@10 -- # set +x 00:05:52.686 ************************************ 00:05:52.686 START TEST event_reactor_perf 00:05:52.686 ************************************ 00:05:52.686 05:33:25 event.event_reactor_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:05:52.686 [2024-12-07 05:33:25.748904] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:52.687 [2024-12-07 05:33:25.749044] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69875 ] 00:05:52.687 [2024-12-07 05:33:25.902688] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:52.687 [2024-12-07 05:33:25.926660] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:53.626 test_start 00:05:53.626 test_end 00:05:53.626 Performance: 399623 events per second 00:05:53.626 00:05:53.626 real 0m1.279s 00:05:53.626 user 0m1.094s 00:05:53.626 sys 0m0.078s 00:05:53.626 05:33:26 event.event_reactor_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:53.626 05:33:26 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:05:53.626 ************************************ 00:05:53.626 END TEST event_reactor_perf 00:05:53.626 ************************************ 00:05:53.886 05:33:27 event -- event/event.sh@49 -- # uname -s 00:05:53.886 05:33:27 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:05:53.886 05:33:27 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:05:53.886 05:33:27 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:53.886 05:33:27 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:53.886 05:33:27 event -- common/autotest_common.sh@10 -- # set +x 00:05:53.886 ************************************ 00:05:53.886 START TEST event_scheduler 00:05:53.886 ************************************ 00:05:53.886 05:33:27 event.event_scheduler -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:05:53.886 * Looking for test storage... 00:05:53.886 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:05:53.886 05:33:27 event.event_scheduler -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:53.886 05:33:27 event.event_scheduler -- common/autotest_common.sh@1711 -- # lcov --version 00:05:53.886 05:33:27 event.event_scheduler -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:54.147 05:33:27 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:54.147 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:54.147 --rc genhtml_branch_coverage=1 00:05:54.147 --rc genhtml_function_coverage=1 00:05:54.147 --rc genhtml_legend=1 00:05:54.147 --rc geninfo_all_blocks=1 00:05:54.147 --rc geninfo_unexecuted_blocks=1 00:05:54.147 00:05:54.147 ' 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:54.147 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:54.147 --rc genhtml_branch_coverage=1 00:05:54.147 --rc genhtml_function_coverage=1 00:05:54.147 --rc genhtml_legend=1 00:05:54.147 --rc geninfo_all_blocks=1 00:05:54.147 --rc geninfo_unexecuted_blocks=1 00:05:54.147 00:05:54.147 ' 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:54.147 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:54.147 --rc genhtml_branch_coverage=1 00:05:54.147 --rc genhtml_function_coverage=1 00:05:54.147 --rc genhtml_legend=1 00:05:54.147 --rc geninfo_all_blocks=1 00:05:54.147 --rc geninfo_unexecuted_blocks=1 00:05:54.147 00:05:54.147 ' 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:54.147 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:54.147 --rc genhtml_branch_coverage=1 00:05:54.147 --rc genhtml_function_coverage=1 00:05:54.147 --rc genhtml_legend=1 00:05:54.147 --rc geninfo_all_blocks=1 00:05:54.147 --rc geninfo_unexecuted_blocks=1 00:05:54.147 00:05:54.147 ' 00:05:54.147 05:33:27 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:05:54.147 05:33:27 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=69945 00:05:54.147 05:33:27 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:05:54.147 05:33:27 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:05:54.147 05:33:27 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 69945 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@835 -- # '[' -z 69945 ']' 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:54.147 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:54.147 05:33:27 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:54.147 [2024-12-07 05:33:27.354920] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:54.147 [2024-12-07 05:33:27.355142] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69945 ] 00:05:54.147 [2024-12-07 05:33:27.499030] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:05:54.408 [2024-12-07 05:33:27.527714] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:54.408 [2024-12-07 05:33:27.527939] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:05:54.408 [2024-12-07 05:33:27.527892] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:54.408 [2024-12-07 05:33:27.528066] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:05:54.978 05:33:28 event.event_scheduler -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:54.978 05:33:28 event.event_scheduler -- common/autotest_common.sh@868 -- # return 0 00:05:54.978 05:33:28 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:05:54.978 05:33:28 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:54.978 05:33:28 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:54.978 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:54.978 POWER: Cannot set governor of lcore 0 to userspace 00:05:54.978 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:54.978 POWER: Cannot set governor of lcore 0 to performance 00:05:54.978 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:54.978 POWER: Cannot set governor of lcore 0 to userspace 00:05:54.978 GUEST_CHANNEL: Unable to to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:05:54.978 POWER: Unable to set Power Management Environment for lcore 0 00:05:54.978 [2024-12-07 05:33:28.188600] dpdk_governor.c: 135:_init_core: *ERROR*: Failed to initialize on core0 00:05:54.978 [2024-12-07 05:33:28.188658] dpdk_governor.c: 196:_init: *ERROR*: Failed to initialize on core0 00:05:54.978 [2024-12-07 05:33:28.188707] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:05:54.978 [2024-12-07 05:33:28.188766] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:05:54.978 [2024-12-07 05:33:28.188799] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:05:54.978 [2024-12-07 05:33:28.188847] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:05:54.978 05:33:28 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:54.978 05:33:28 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:05:54.978 05:33:28 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:54.978 05:33:28 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:54.978 [2024-12-07 05:33:28.266271] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:05:54.978 05:33:28 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:54.978 05:33:28 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:05:54.978 05:33:28 event.event_scheduler -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:54.978 05:33:28 event.event_scheduler -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:54.978 05:33:28 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:54.978 ************************************ 00:05:54.978 START TEST scheduler_create_thread 00:05:54.978 ************************************ 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1129 -- # scheduler_create_thread 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:54.978 2 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:54.978 3 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:54.978 4 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:54.978 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:54.978 5 00:05:54.979 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:54.979 05:33:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:05:54.979 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:54.979 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:55.239 6 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:55.239 7 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:55.239 8 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:55.239 9 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:55.239 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:55.499 10 00:05:55.499 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:55.499 05:33:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:05:55.499 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:55.499 05:33:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:56.881 05:33:30 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:56.881 05:33:30 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:05:56.881 05:33:30 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:05:56.881 05:33:30 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:56.881 05:33:30 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:57.822 05:33:30 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:57.822 05:33:30 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:05:57.822 05:33:30 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:57.822 05:33:30 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:58.392 05:33:31 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:58.392 05:33:31 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:05:58.392 05:33:31 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:05:58.392 05:33:31 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:58.392 05:33:31 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.331 ************************************ 00:05:59.331 END TEST scheduler_create_thread 00:05:59.331 ************************************ 00:05:59.331 05:33:32 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:59.331 00:05:59.331 real 0m4.211s 00:05:59.331 user 0m0.026s 00:05:59.331 sys 0m0.010s 00:05:59.331 05:33:32 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:59.331 05:33:32 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.331 05:33:32 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:05:59.331 05:33:32 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 69945 00:05:59.331 05:33:32 event.event_scheduler -- common/autotest_common.sh@954 -- # '[' -z 69945 ']' 00:05:59.331 05:33:32 event.event_scheduler -- common/autotest_common.sh@958 -- # kill -0 69945 00:05:59.331 05:33:32 event.event_scheduler -- common/autotest_common.sh@959 -- # uname 00:05:59.331 05:33:32 event.event_scheduler -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:59.331 05:33:32 event.event_scheduler -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69945 00:05:59.331 killing process with pid 69945 00:05:59.331 05:33:32 event.event_scheduler -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:05:59.331 05:33:32 event.event_scheduler -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:05:59.331 05:33:32 event.event_scheduler -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69945' 00:05:59.331 05:33:32 event.event_scheduler -- common/autotest_common.sh@973 -- # kill 69945 00:05:59.331 05:33:32 event.event_scheduler -- common/autotest_common.sh@978 -- # wait 69945 00:05:59.591 [2024-12-07 05:33:32.770328] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:05:59.851 00:05:59.851 real 0m5.974s 00:05:59.851 user 0m12.995s 00:05:59.851 sys 0m0.463s 00:05:59.851 ************************************ 00:05:59.851 END TEST event_scheduler 00:05:59.851 ************************************ 00:05:59.851 05:33:33 event.event_scheduler -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:59.851 05:33:33 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:59.851 05:33:33 event -- event/event.sh@51 -- # modprobe -n nbd 00:05:59.851 05:33:33 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:05:59.851 05:33:33 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:59.851 05:33:33 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:59.851 05:33:33 event -- common/autotest_common.sh@10 -- # set +x 00:05:59.851 ************************************ 00:05:59.851 START TEST app_repeat 00:05:59.851 ************************************ 00:05:59.851 05:33:33 event.app_repeat -- common/autotest_common.sh@1129 -- # app_repeat_test 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@19 -- # repeat_pid=70058 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 70058' 00:05:59.851 Process app_repeat pid: 70058 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:05:59.851 spdk_app_start Round 0 00:05:59.851 05:33:33 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70058 /var/tmp/spdk-nbd.sock 00:05:59.851 05:33:33 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70058 ']' 00:05:59.851 05:33:33 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:05:59.851 05:33:33 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:59.851 05:33:33 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:05:59.851 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:05:59.851 05:33:33 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:59.851 05:33:33 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:05:59.851 [2024-12-07 05:33:33.162305] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:05:59.851 [2024-12-07 05:33:33.162490] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70058 ] 00:06:00.110 [2024-12-07 05:33:33.316330] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:00.110 [2024-12-07 05:33:33.342786] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:00.110 [2024-12-07 05:33:33.342882] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:00.681 05:33:34 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:00.681 05:33:34 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:00.681 05:33:34 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:00.941 Malloc0 00:06:00.941 05:33:34 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:01.201 Malloc1 00:06:01.201 05:33:34 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:01.201 05:33:34 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:01.461 /dev/nbd0 00:06:01.461 05:33:34 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:01.461 05:33:34 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:01.461 1+0 records in 00:06:01.461 1+0 records out 00:06:01.461 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000330648 s, 12.4 MB/s 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:01.461 05:33:34 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:01.461 05:33:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:01.461 05:33:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:01.461 05:33:34 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:01.721 /dev/nbd1 00:06:01.721 05:33:34 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:01.721 05:33:34 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:01.721 1+0 records in 00:06:01.721 1+0 records out 00:06:01.721 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00040354 s, 10.2 MB/s 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:01.721 05:33:34 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:01.721 05:33:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:01.721 05:33:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:01.721 05:33:34 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:01.721 05:33:34 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:01.721 05:33:34 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:01.980 { 00:06:01.980 "nbd_device": "/dev/nbd0", 00:06:01.980 "bdev_name": "Malloc0" 00:06:01.980 }, 00:06:01.980 { 00:06:01.980 "nbd_device": "/dev/nbd1", 00:06:01.980 "bdev_name": "Malloc1" 00:06:01.980 } 00:06:01.980 ]' 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:01.980 { 00:06:01.980 "nbd_device": "/dev/nbd0", 00:06:01.980 "bdev_name": "Malloc0" 00:06:01.980 }, 00:06:01.980 { 00:06:01.980 "nbd_device": "/dev/nbd1", 00:06:01.980 "bdev_name": "Malloc1" 00:06:01.980 } 00:06:01.980 ]' 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:01.980 /dev/nbd1' 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:01.980 /dev/nbd1' 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:01.980 05:33:35 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:01.981 256+0 records in 00:06:01.981 256+0 records out 00:06:01.981 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0125306 s, 83.7 MB/s 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:01.981 256+0 records in 00:06:01.981 256+0 records out 00:06:01.981 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0234045 s, 44.8 MB/s 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:01.981 256+0 records in 00:06:01.981 256+0 records out 00:06:01.981 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0276205 s, 38.0 MB/s 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:01.981 05:33:35 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:02.240 05:33:35 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:02.240 05:33:35 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:02.240 05:33:35 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:02.240 05:33:35 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:02.240 05:33:35 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:02.240 05:33:35 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:02.240 05:33:35 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:02.240 05:33:35 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:02.240 05:33:35 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:02.240 05:33:35 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:02.498 05:33:35 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:02.498 05:33:35 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:02.498 05:33:35 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:02.498 05:33:35 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:02.498 05:33:35 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:02.498 05:33:35 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:02.498 05:33:35 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:02.498 05:33:35 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:02.498 05:33:35 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:02.498 05:33:35 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:02.498 05:33:35 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:02.758 05:33:35 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:02.758 05:33:35 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:03.018 05:33:36 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:03.018 [2024-12-07 05:33:36.356961] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:03.018 [2024-12-07 05:33:36.380027] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:03.018 [2024-12-07 05:33:36.380035] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:03.279 [2024-12-07 05:33:36.420899] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:03.279 [2024-12-07 05:33:36.420976] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:06.610 spdk_app_start Round 1 00:06:06.610 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:06.610 05:33:39 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:06.610 05:33:39 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:06:06.610 05:33:39 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70058 /var/tmp/spdk-nbd.sock 00:06:06.610 05:33:39 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70058 ']' 00:06:06.610 05:33:39 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:06.610 05:33:39 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:06.610 05:33:39 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:06.610 05:33:39 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:06.610 05:33:39 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:06.610 05:33:39 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:06.610 05:33:39 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:06.610 05:33:39 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:06.610 Malloc0 00:06:06.610 05:33:39 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:06.610 Malloc1 00:06:06.610 05:33:39 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:06.610 05:33:39 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:06.869 /dev/nbd0 00:06:06.869 05:33:40 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:06.869 05:33:40 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:06.869 1+0 records in 00:06:06.869 1+0 records out 00:06:06.869 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000359067 s, 11.4 MB/s 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:06.869 05:33:40 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:06.869 05:33:40 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:06.869 05:33:40 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:06.869 05:33:40 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:07.129 /dev/nbd1 00:06:07.129 05:33:40 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:07.129 05:33:40 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:07.129 1+0 records in 00:06:07.129 1+0 records out 00:06:07.129 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000236281 s, 17.3 MB/s 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:07.129 05:33:40 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:07.129 05:33:40 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:07.129 05:33:40 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:07.129 05:33:40 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:07.129 05:33:40 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:07.129 05:33:40 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:07.389 { 00:06:07.389 "nbd_device": "/dev/nbd0", 00:06:07.389 "bdev_name": "Malloc0" 00:06:07.389 }, 00:06:07.389 { 00:06:07.389 "nbd_device": "/dev/nbd1", 00:06:07.389 "bdev_name": "Malloc1" 00:06:07.389 } 00:06:07.389 ]' 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:07.389 { 00:06:07.389 "nbd_device": "/dev/nbd0", 00:06:07.389 "bdev_name": "Malloc0" 00:06:07.389 }, 00:06:07.389 { 00:06:07.389 "nbd_device": "/dev/nbd1", 00:06:07.389 "bdev_name": "Malloc1" 00:06:07.389 } 00:06:07.389 ]' 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:07.389 /dev/nbd1' 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:07.389 /dev/nbd1' 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:07.389 256+0 records in 00:06:07.389 256+0 records out 00:06:07.389 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0127231 s, 82.4 MB/s 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:07.389 256+0 records in 00:06:07.389 256+0 records out 00:06:07.389 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0203899 s, 51.4 MB/s 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:07.389 256+0 records in 00:06:07.389 256+0 records out 00:06:07.389 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0220182 s, 47.6 MB/s 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:07.389 05:33:40 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:07.649 05:33:40 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:07.649 05:33:40 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:07.649 05:33:40 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:07.649 05:33:40 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:07.649 05:33:40 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:07.649 05:33:40 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:07.649 05:33:40 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:07.649 05:33:40 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:07.649 05:33:40 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:07.649 05:33:40 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:07.908 05:33:41 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:07.908 05:33:41 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:07.908 05:33:41 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:07.908 05:33:41 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:07.908 05:33:41 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:07.908 05:33:41 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:07.908 05:33:41 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:07.908 05:33:41 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:07.908 05:33:41 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:07.908 05:33:41 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:07.908 05:33:41 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:08.168 05:33:41 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:08.168 05:33:41 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:08.427 05:33:41 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:08.685 [2024-12-07 05:33:41.805804] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:08.685 [2024-12-07 05:33:41.828691] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:08.685 [2024-12-07 05:33:41.828716] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:08.685 [2024-12-07 05:33:41.869892] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:08.685 [2024-12-07 05:33:41.869967] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:11.976 spdk_app_start Round 2 00:06:11.976 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:11.976 05:33:44 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:11.976 05:33:44 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:06:11.976 05:33:44 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70058 /var/tmp/spdk-nbd.sock 00:06:11.976 05:33:44 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70058 ']' 00:06:11.976 05:33:44 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:11.976 05:33:44 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:11.976 05:33:44 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:11.976 05:33:44 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:11.976 05:33:44 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:11.976 05:33:44 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:11.976 05:33:44 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:11.976 05:33:44 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:11.976 Malloc0 00:06:11.976 05:33:45 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:11.976 Malloc1 00:06:11.976 05:33:45 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:11.976 05:33:45 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:12.235 /dev/nbd0 00:06:12.235 05:33:45 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:12.235 05:33:45 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:12.235 1+0 records in 00:06:12.235 1+0 records out 00:06:12.235 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000360672 s, 11.4 MB/s 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:12.235 05:33:45 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:12.235 05:33:45 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:12.235 05:33:45 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:12.235 05:33:45 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:12.493 /dev/nbd1 00:06:12.493 05:33:45 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:12.493 05:33:45 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:12.493 05:33:45 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:06:12.493 05:33:45 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:12.493 05:33:45 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:12.493 05:33:45 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:12.493 05:33:45 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:06:12.494 05:33:45 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:12.494 05:33:45 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:12.494 05:33:45 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:12.494 05:33:45 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:12.494 1+0 records in 00:06:12.494 1+0 records out 00:06:12.494 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000383391 s, 10.7 MB/s 00:06:12.494 05:33:45 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:12.494 05:33:45 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:12.494 05:33:45 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:12.494 05:33:45 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:12.494 05:33:45 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:12.494 05:33:45 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:12.494 05:33:45 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:12.494 05:33:45 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:12.494 05:33:45 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:12.494 05:33:45 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:12.752 05:33:45 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:12.752 { 00:06:12.752 "nbd_device": "/dev/nbd0", 00:06:12.752 "bdev_name": "Malloc0" 00:06:12.752 }, 00:06:12.752 { 00:06:12.752 "nbd_device": "/dev/nbd1", 00:06:12.752 "bdev_name": "Malloc1" 00:06:12.752 } 00:06:12.752 ]' 00:06:12.752 05:33:45 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:12.752 { 00:06:12.752 "nbd_device": "/dev/nbd0", 00:06:12.752 "bdev_name": "Malloc0" 00:06:12.752 }, 00:06:12.752 { 00:06:12.752 "nbd_device": "/dev/nbd1", 00:06:12.752 "bdev_name": "Malloc1" 00:06:12.752 } 00:06:12.752 ]' 00:06:12.752 05:33:45 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:12.752 /dev/nbd1' 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:12.752 /dev/nbd1' 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:12.752 256+0 records in 00:06:12.752 256+0 records out 00:06:12.752 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0051804 s, 202 MB/s 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:12.752 256+0 records in 00:06:12.752 256+0 records out 00:06:12.752 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0200364 s, 52.3 MB/s 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:12.752 256+0 records in 00:06:12.752 256+0 records out 00:06:12.752 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0211074 s, 49.7 MB/s 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:12.752 05:33:46 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:13.010 05:33:46 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:13.269 05:33:46 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:13.269 05:33:46 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:13.269 05:33:46 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:13.269 05:33:46 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:13.269 05:33:46 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:13.269 05:33:46 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:13.269 05:33:46 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:13.269 05:33:46 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:13.269 05:33:46 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:13.269 05:33:46 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:13.269 05:33:46 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:13.528 05:33:46 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:13.528 05:33:46 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:13.788 05:33:47 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:13.788 [2024-12-07 05:33:47.136298] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:14.048 [2024-12-07 05:33:47.159196] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:14.048 [2024-12-07 05:33:47.159196] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:14.048 [2024-12-07 05:33:47.200205] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:14.048 [2024-12-07 05:33:47.200263] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:17.343 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:17.343 05:33:50 event.app_repeat -- event/event.sh@38 -- # waitforlisten 70058 /var/tmp/spdk-nbd.sock 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70058 ']' 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:17.343 05:33:50 event.app_repeat -- event/event.sh@39 -- # killprocess 70058 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@954 -- # '[' -z 70058 ']' 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@958 -- # kill -0 70058 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@959 -- # uname 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70058 00:06:17.343 killing process with pid 70058 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70058' 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@973 -- # kill 70058 00:06:17.343 05:33:50 event.app_repeat -- common/autotest_common.sh@978 -- # wait 70058 00:06:17.343 spdk_app_start is called in Round 0. 00:06:17.343 Shutdown signal received, stop current app iteration 00:06:17.343 Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 reinitialization... 00:06:17.343 spdk_app_start is called in Round 1. 00:06:17.343 Shutdown signal received, stop current app iteration 00:06:17.343 Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 reinitialization... 00:06:17.343 spdk_app_start is called in Round 2. 00:06:17.343 Shutdown signal received, stop current app iteration 00:06:17.343 Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 reinitialization... 00:06:17.343 spdk_app_start is called in Round 3. 00:06:17.343 Shutdown signal received, stop current app iteration 00:06:17.343 05:33:50 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:06:17.344 05:33:50 event.app_repeat -- event/event.sh@42 -- # return 0 00:06:17.344 00:06:17.344 real 0m17.312s 00:06:17.344 user 0m38.280s 00:06:17.344 sys 0m2.628s 00:06:17.344 05:33:50 event.app_repeat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:17.344 05:33:50 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:17.344 ************************************ 00:06:17.344 END TEST app_repeat 00:06:17.344 ************************************ 00:06:17.344 05:33:50 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:06:17.344 05:33:50 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:17.344 05:33:50 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:17.344 05:33:50 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:17.344 05:33:50 event -- common/autotest_common.sh@10 -- # set +x 00:06:17.344 ************************************ 00:06:17.344 START TEST cpu_locks 00:06:17.344 ************************************ 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:17.344 * Looking for test storage... 00:06:17.344 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1711 -- # lcov --version 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:17.344 05:33:50 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:17.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:17.344 --rc genhtml_branch_coverage=1 00:06:17.344 --rc genhtml_function_coverage=1 00:06:17.344 --rc genhtml_legend=1 00:06:17.344 --rc geninfo_all_blocks=1 00:06:17.344 --rc geninfo_unexecuted_blocks=1 00:06:17.344 00:06:17.344 ' 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:17.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:17.344 --rc genhtml_branch_coverage=1 00:06:17.344 --rc genhtml_function_coverage=1 00:06:17.344 --rc genhtml_legend=1 00:06:17.344 --rc geninfo_all_blocks=1 00:06:17.344 --rc geninfo_unexecuted_blocks=1 00:06:17.344 00:06:17.344 ' 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:17.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:17.344 --rc genhtml_branch_coverage=1 00:06:17.344 --rc genhtml_function_coverage=1 00:06:17.344 --rc genhtml_legend=1 00:06:17.344 --rc geninfo_all_blocks=1 00:06:17.344 --rc geninfo_unexecuted_blocks=1 00:06:17.344 00:06:17.344 ' 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:17.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:17.344 --rc genhtml_branch_coverage=1 00:06:17.344 --rc genhtml_function_coverage=1 00:06:17.344 --rc genhtml_legend=1 00:06:17.344 --rc geninfo_all_blocks=1 00:06:17.344 --rc geninfo_unexecuted_blocks=1 00:06:17.344 00:06:17.344 ' 00:06:17.344 05:33:50 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:06:17.344 05:33:50 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:06:17.344 05:33:50 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:06:17.344 05:33:50 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:17.344 05:33:50 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:17.605 ************************************ 00:06:17.605 START TEST default_locks 00:06:17.605 ************************************ 00:06:17.605 05:33:50 event.cpu_locks.default_locks -- common/autotest_common.sh@1129 -- # default_locks 00:06:17.605 05:33:50 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=70482 00:06:17.605 05:33:50 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:17.605 05:33:50 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 70482 00:06:17.605 05:33:50 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 70482 ']' 00:06:17.605 05:33:50 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:17.605 05:33:50 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:17.605 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:17.605 05:33:50 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:17.605 05:33:50 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:17.605 05:33:50 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:17.605 [2024-12-07 05:33:50.801681] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:17.605 [2024-12-07 05:33:50.801830] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70482 ] 00:06:17.605 [2024-12-07 05:33:50.957706] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:17.864 [2024-12-07 05:33:50.982407] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:18.433 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:18.433 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 0 00:06:18.433 05:33:51 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 70482 00:06:18.433 05:33:51 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 70482 00:06:18.433 05:33:51 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:18.705 05:33:51 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 70482 00:06:18.705 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # '[' -z 70482 ']' 00:06:18.705 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # kill -0 70482 00:06:18.705 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # uname 00:06:18.705 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:18.705 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70482 00:06:18.705 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:18.705 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:18.705 killing process with pid 70482 00:06:18.705 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70482' 00:06:18.705 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@973 -- # kill 70482 00:06:18.705 05:33:51 event.cpu_locks.default_locks -- common/autotest_common.sh@978 -- # wait 70482 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 70482 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # local es=0 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70482 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # waitforlisten 70482 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 70482 ']' 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:18.979 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:18.979 ERROR: process (pid: 70482) is no longer running 00:06:18.979 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70482) - No such process 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 1 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # es=1 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:18.979 00:06:18.979 real 0m1.540s 00:06:18.979 user 0m1.516s 00:06:18.979 sys 0m0.501s 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:18.979 05:33:52 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:18.979 ************************************ 00:06:18.979 END TEST default_locks 00:06:18.979 ************************************ 00:06:18.979 05:33:52 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:06:18.979 05:33:52 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:18.979 05:33:52 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:18.979 05:33:52 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:18.979 ************************************ 00:06:18.979 START TEST default_locks_via_rpc 00:06:18.979 ************************************ 00:06:18.979 05:33:52 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1129 -- # default_locks_via_rpc 00:06:18.979 05:33:52 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=70529 00:06:18.979 05:33:52 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:18.979 05:33:52 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 70529 00:06:18.979 05:33:52 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70529 ']' 00:06:18.979 05:33:52 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:18.980 05:33:52 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:18.980 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:18.980 05:33:52 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:18.980 05:33:52 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:18.980 05:33:52 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:19.239 [2024-12-07 05:33:52.414882] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:19.239 [2024-12-07 05:33:52.415036] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70529 ] 00:06:19.239 [2024-12-07 05:33:52.567998] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:19.239 [2024-12-07 05:33:52.592956] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 70529 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 70529 00:06:20.180 05:33:53 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:20.441 05:33:53 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 70529 00:06:20.441 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # '[' -z 70529 ']' 00:06:20.441 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # kill -0 70529 00:06:20.441 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # uname 00:06:20.441 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:20.441 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70529 00:06:20.441 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:20.441 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:20.441 killing process with pid 70529 00:06:20.441 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70529' 00:06:20.441 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@973 -- # kill 70529 00:06:20.441 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@978 -- # wait 70529 00:06:20.701 00:06:20.701 real 0m1.619s 00:06:20.701 user 0m1.617s 00:06:20.701 sys 0m0.525s 00:06:20.701 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:20.701 05:33:53 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:20.701 ************************************ 00:06:20.701 END TEST default_locks_via_rpc 00:06:20.701 ************************************ 00:06:20.701 05:33:53 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:06:20.701 05:33:53 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:20.701 05:33:53 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:20.701 05:33:53 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:20.701 ************************************ 00:06:20.701 START TEST non_locking_app_on_locked_coremask 00:06:20.701 ************************************ 00:06:20.701 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # non_locking_app_on_locked_coremask 00:06:20.701 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=70576 00:06:20.701 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:20.701 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 70576 /var/tmp/spdk.sock 00:06:20.701 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70576 ']' 00:06:20.701 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:20.701 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:20.701 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:20.701 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:20.701 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:20.701 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:20.961 [2024-12-07 05:33:54.101138] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:20.961 [2024-12-07 05:33:54.101265] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70576 ] 00:06:20.961 [2024-12-07 05:33:54.254711] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:20.961 [2024-12-07 05:33:54.279269] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:21.903 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:21.903 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:21.903 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=70592 00:06:21.903 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:06:21.903 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 70592 /var/tmp/spdk2.sock 00:06:21.903 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70592 ']' 00:06:21.903 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:21.903 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:21.903 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:21.903 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:21.903 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:21.903 05:33:54 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:21.903 [2024-12-07 05:33:55.002702] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:21.903 [2024-12-07 05:33:55.002850] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70592 ] 00:06:21.903 [2024-12-07 05:33:55.150826] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:21.903 [2024-12-07 05:33:55.150899] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:21.903 [2024-12-07 05:33:55.204049] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:22.472 05:33:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:22.472 05:33:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:22.472 05:33:55 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 70576 00:06:22.472 05:33:55 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70576 00:06:22.472 05:33:55 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:23.041 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 70576 00:06:23.041 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70576 ']' 00:06:23.041 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70576 00:06:23.041 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:23.041 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:23.041 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70576 00:06:23.041 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:23.041 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:23.041 killing process with pid 70576 00:06:23.041 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70576' 00:06:23.041 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70576 00:06:23.041 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70576 00:06:23.610 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 70592 00:06:23.610 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70592 ']' 00:06:23.610 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70592 00:06:23.610 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:23.610 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:23.870 05:33:56 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70592 00:06:23.870 05:33:57 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:23.870 05:33:57 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:23.870 killing process with pid 70592 00:06:23.870 05:33:57 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70592' 00:06:23.870 05:33:57 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70592 00:06:23.870 05:33:57 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70592 00:06:24.130 00:06:24.130 real 0m3.358s 00:06:24.130 user 0m3.533s 00:06:24.130 sys 0m0.995s 00:06:24.130 05:33:57 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:24.130 ************************************ 00:06:24.130 END TEST non_locking_app_on_locked_coremask 00:06:24.130 ************************************ 00:06:24.130 05:33:57 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:24.130 05:33:57 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:06:24.130 05:33:57 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:24.130 05:33:57 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:24.130 05:33:57 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:24.130 ************************************ 00:06:24.130 START TEST locking_app_on_unlocked_coremask 00:06:24.130 ************************************ 00:06:24.130 05:33:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_unlocked_coremask 00:06:24.130 05:33:57 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=70655 00:06:24.130 05:33:57 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 70655 /var/tmp/spdk.sock 00:06:24.130 05:33:57 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:06:24.130 05:33:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70655 ']' 00:06:24.130 05:33:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:24.130 05:33:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:24.130 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:24.130 05:33:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:24.130 05:33:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:24.130 05:33:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:24.390 [2024-12-07 05:33:57.521507] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:24.390 [2024-12-07 05:33:57.521677] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70655 ] 00:06:24.390 [2024-12-07 05:33:57.674922] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:24.390 [2024-12-07 05:33:57.674983] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:24.390 [2024-12-07 05:33:57.699457] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:24.958 05:33:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:24.958 05:33:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:24.958 05:33:58 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:24.958 05:33:58 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=70666 00:06:25.217 05:33:58 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 70666 /var/tmp/spdk2.sock 00:06:25.217 05:33:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70666 ']' 00:06:25.217 05:33:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:25.217 05:33:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:25.217 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:25.217 05:33:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:25.217 05:33:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:25.217 05:33:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:25.217 [2024-12-07 05:33:58.391920] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:25.217 [2024-12-07 05:33:58.392047] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70666 ] 00:06:25.217 [2024-12-07 05:33:58.543741] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:25.476 [2024-12-07 05:33:58.596260] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:26.044 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:26.044 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:26.044 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 70666 00:06:26.044 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70666 00:06:26.044 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:26.303 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 70655 00:06:26.303 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70655 ']' 00:06:26.303 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 70655 00:06:26.303 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:26.303 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:26.303 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70655 00:06:26.303 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:26.303 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:26.303 killing process with pid 70655 00:06:26.303 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70655' 00:06:26.303 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 70655 00:06:26.303 05:33:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 70655 00:06:27.242 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 70666 00:06:27.242 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70666 ']' 00:06:27.242 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 70666 00:06:27.242 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:27.242 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:27.242 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70666 00:06:27.242 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:27.242 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:27.242 killing process with pid 70666 00:06:27.242 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70666' 00:06:27.242 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 70666 00:06:27.242 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 70666 00:06:27.501 00:06:27.501 real 0m3.296s 00:06:27.501 user 0m3.452s 00:06:27.501 sys 0m0.971s 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:27.501 ************************************ 00:06:27.501 END TEST locking_app_on_unlocked_coremask 00:06:27.501 ************************************ 00:06:27.501 05:34:00 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:06:27.501 05:34:00 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:27.501 05:34:00 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:27.501 05:34:00 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:27.501 ************************************ 00:06:27.501 START TEST locking_app_on_locked_coremask 00:06:27.501 ************************************ 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_locked_coremask 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=70737 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 70737 /var/tmp/spdk.sock 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70737 ']' 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:27.501 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:27.501 05:34:00 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:27.760 [2024-12-07 05:34:00.881328] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:27.760 [2024-12-07 05:34:00.881458] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70737 ] 00:06:27.760 [2024-12-07 05:34:01.030121] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:27.760 [2024-12-07 05:34:01.054569] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=70747 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 70747 /var/tmp/spdk2.sock 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # local es=0 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70747 /var/tmp/spdk2.sock 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # waitforlisten 70747 /var/tmp/spdk2.sock 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70747 ']' 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:28.707 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:28.707 05:34:01 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:28.707 [2024-12-07 05:34:01.802298] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:28.708 [2024-12-07 05:34:01.802503] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70747 ] 00:06:28.708 [2024-12-07 05:34:01.959977] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 70737 has claimed it. 00:06:28.708 [2024-12-07 05:34:01.960042] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:29.290 ERROR: process (pid: 70747) is no longer running 00:06:29.290 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70747) - No such process 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 1 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # es=1 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 70737 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70737 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 70737 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70737 ']' 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70737 00:06:29.290 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:29.549 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:29.549 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70737 00:06:29.549 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:29.549 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:29.549 killing process with pid 70737 00:06:29.549 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70737' 00:06:29.549 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70737 00:06:29.549 05:34:02 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70737 00:06:29.808 00:06:29.808 real 0m2.257s 00:06:29.808 user 0m2.450s 00:06:29.808 sys 0m0.645s 00:06:29.808 05:34:03 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:29.808 05:34:03 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:29.808 ************************************ 00:06:29.808 END TEST locking_app_on_locked_coremask 00:06:29.808 ************************************ 00:06:29.808 05:34:03 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:06:29.808 05:34:03 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:29.808 05:34:03 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:29.808 05:34:03 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:29.808 ************************************ 00:06:29.808 START TEST locking_overlapped_coremask 00:06:29.808 ************************************ 00:06:29.808 05:34:03 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask 00:06:29.808 05:34:03 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=70795 00:06:29.808 05:34:03 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:06:29.808 05:34:03 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 70795 /var/tmp/spdk.sock 00:06:29.808 05:34:03 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 70795 ']' 00:06:29.808 05:34:03 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:29.808 05:34:03 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:29.808 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:29.808 05:34:03 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:29.808 05:34:03 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:29.808 05:34:03 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:30.068 [2024-12-07 05:34:03.203911] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:30.068 [2024-12-07 05:34:03.204051] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70795 ] 00:06:30.068 [2024-12-07 05:34:03.357993] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:30.068 [2024-12-07 05:34:03.385109] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:30.068 [2024-12-07 05:34:03.385192] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:30.068 [2024-12-07 05:34:03.385309] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=70813 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 70813 /var/tmp/spdk2.sock 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # local es=0 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70813 /var/tmp/spdk2.sock 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # waitforlisten 70813 /var/tmp/spdk2.sock 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 70813 ']' 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:31.008 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:31.008 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:31.008 [2024-12-07 05:34:04.086528] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:31.008 [2024-12-07 05:34:04.086664] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70813 ] 00:06:31.008 [2024-12-07 05:34:04.238696] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 70795 has claimed it. 00:06:31.008 [2024-12-07 05:34:04.238757] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:31.578 ERROR: process (pid: 70813) is no longer running 00:06:31.578 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70813) - No such process 00:06:31.578 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:31.578 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 1 00:06:31.578 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # es=1 00:06:31.578 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:31.578 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 70795 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # '[' -z 70795 ']' 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # kill -0 70795 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # uname 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70795 00:06:31.579 killing process with pid 70795 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70795' 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@973 -- # kill 70795 00:06:31.579 05:34:04 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@978 -- # wait 70795 00:06:31.839 ************************************ 00:06:31.839 END TEST locking_overlapped_coremask 00:06:31.839 ************************************ 00:06:31.839 00:06:31.839 real 0m2.033s 00:06:31.839 user 0m5.475s 00:06:31.839 sys 0m0.494s 00:06:31.839 05:34:05 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:31.839 05:34:05 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:31.839 05:34:05 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:06:31.839 05:34:05 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:31.839 05:34:05 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:31.839 05:34:05 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:32.097 ************************************ 00:06:32.097 START TEST locking_overlapped_coremask_via_rpc 00:06:32.097 ************************************ 00:06:32.097 05:34:05 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask_via_rpc 00:06:32.097 05:34:05 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=70855 00:06:32.097 05:34:05 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:06:32.097 05:34:05 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 70855 /var/tmp/spdk.sock 00:06:32.097 05:34:05 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70855 ']' 00:06:32.097 05:34:05 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:32.097 05:34:05 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:32.097 05:34:05 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:32.097 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:32.097 05:34:05 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:32.097 05:34:05 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:32.097 [2024-12-07 05:34:05.304560] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:32.097 [2024-12-07 05:34:05.304820] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70855 ] 00:06:32.097 [2024-12-07 05:34:05.460417] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:32.097 [2024-12-07 05:34:05.460544] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:32.356 [2024-12-07 05:34:05.487162] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:32.356 [2024-12-07 05:34:05.487278] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:32.356 [2024-12-07 05:34:05.487395] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:32.926 05:34:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:32.926 05:34:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:32.926 05:34:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=70873 00:06:32.926 05:34:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:06:32.926 05:34:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 70873 /var/tmp/spdk2.sock 00:06:32.926 05:34:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70873 ']' 00:06:32.926 05:34:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:32.926 05:34:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:32.926 05:34:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:32.926 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:32.926 05:34:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:32.926 05:34:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:32.926 [2024-12-07 05:34:06.208991] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:32.926 [2024-12-07 05:34:06.209218] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70873 ] 00:06:33.186 [2024-12-07 05:34:06.359162] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:33.186 [2024-12-07 05:34:06.359227] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:33.186 [2024-12-07 05:34:06.419868] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:33.186 [2024-12-07 05:34:06.419789] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:33.186 [2024-12-07 05:34:06.420013] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # local es=0 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:33.757 [2024-12-07 05:34:07.057792] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 70855 has claimed it. 00:06:33.757 request: 00:06:33.757 { 00:06:33.757 "method": "framework_enable_cpumask_locks", 00:06:33.757 "req_id": 1 00:06:33.757 } 00:06:33.757 Got JSON-RPC error response 00:06:33.757 response: 00:06:33.757 { 00:06:33.757 "code": -32603, 00:06:33.757 "message": "Failed to claim CPU core: 2" 00:06:33.757 } 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # es=1 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 70855 /var/tmp/spdk.sock 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70855 ']' 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:33.757 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:33.757 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:34.017 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:34.017 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:34.017 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 70873 /var/tmp/spdk2.sock 00:06:34.017 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70873 ']' 00:06:34.017 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:34.017 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:34.017 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:34.017 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:34.017 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:34.017 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:34.277 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:34.277 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:34.277 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:06:34.277 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:34.277 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:34.277 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:34.277 00:06:34.277 real 0m2.280s 00:06:34.277 user 0m1.049s 00:06:34.277 sys 0m0.162s 00:06:34.277 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:34.277 05:34:07 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:34.277 ************************************ 00:06:34.277 END TEST locking_overlapped_coremask_via_rpc 00:06:34.277 ************************************ 00:06:34.277 05:34:07 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:06:34.277 05:34:07 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 70855 ]] 00:06:34.277 05:34:07 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 70855 00:06:34.277 05:34:07 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70855 ']' 00:06:34.277 05:34:07 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70855 00:06:34.277 05:34:07 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:06:34.277 05:34:07 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:34.277 05:34:07 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70855 00:06:34.277 05:34:07 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:34.277 killing process with pid 70855 00:06:34.277 05:34:07 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:34.277 05:34:07 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70855' 00:06:34.277 05:34:07 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 70855 00:06:34.277 05:34:07 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 70855 00:06:34.848 05:34:07 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 70873 ]] 00:06:34.848 05:34:07 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 70873 00:06:34.848 05:34:07 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70873 ']' 00:06:34.848 05:34:07 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70873 00:06:34.848 05:34:07 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:06:34.848 05:34:07 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:34.848 05:34:07 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70873 00:06:34.848 05:34:08 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:06:34.848 05:34:08 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:06:34.848 killing process with pid 70873 00:06:34.848 05:34:08 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70873' 00:06:34.848 05:34:08 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 70873 00:06:34.848 05:34:08 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 70873 00:06:35.106 05:34:08 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:35.107 Process with pid 70855 is not found 00:06:35.107 05:34:08 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:06:35.107 05:34:08 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 70855 ]] 00:06:35.107 05:34:08 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 70855 00:06:35.107 05:34:08 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70855 ']' 00:06:35.107 05:34:08 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70855 00:06:35.107 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (70855) - No such process 00:06:35.107 05:34:08 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 70855 is not found' 00:06:35.107 05:34:08 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 70873 ]] 00:06:35.107 05:34:08 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 70873 00:06:35.107 05:34:08 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70873 ']' 00:06:35.107 05:34:08 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70873 00:06:35.107 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (70873) - No such process 00:06:35.107 05:34:08 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 70873 is not found' 00:06:35.107 Process with pid 70873 is not found 00:06:35.107 05:34:08 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:35.107 00:06:35.107 real 0m17.899s 00:06:35.107 user 0m30.481s 00:06:35.107 sys 0m5.326s 00:06:35.107 05:34:08 event.cpu_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:35.107 05:34:08 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:35.107 ************************************ 00:06:35.107 END TEST cpu_locks 00:06:35.107 ************************************ 00:06:35.107 00:06:35.107 real 0m45.644s 00:06:35.107 user 1m28.273s 00:06:35.107 sys 0m9.026s 00:06:35.107 05:34:08 event -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:35.107 05:34:08 event -- common/autotest_common.sh@10 -- # set +x 00:06:35.107 ************************************ 00:06:35.107 END TEST event 00:06:35.107 ************************************ 00:06:35.365 05:34:08 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:35.365 05:34:08 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:35.365 05:34:08 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:35.365 05:34:08 -- common/autotest_common.sh@10 -- # set +x 00:06:35.365 ************************************ 00:06:35.365 START TEST thread 00:06:35.365 ************************************ 00:06:35.365 05:34:08 thread -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:35.365 * Looking for test storage... 00:06:35.365 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:06:35.365 05:34:08 thread -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:35.365 05:34:08 thread -- common/autotest_common.sh@1711 -- # lcov --version 00:06:35.365 05:34:08 thread -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:35.365 05:34:08 thread -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:35.365 05:34:08 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:35.366 05:34:08 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:35.366 05:34:08 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:35.366 05:34:08 thread -- scripts/common.sh@336 -- # IFS=.-: 00:06:35.366 05:34:08 thread -- scripts/common.sh@336 -- # read -ra ver1 00:06:35.366 05:34:08 thread -- scripts/common.sh@337 -- # IFS=.-: 00:06:35.366 05:34:08 thread -- scripts/common.sh@337 -- # read -ra ver2 00:06:35.366 05:34:08 thread -- scripts/common.sh@338 -- # local 'op=<' 00:06:35.366 05:34:08 thread -- scripts/common.sh@340 -- # ver1_l=2 00:06:35.366 05:34:08 thread -- scripts/common.sh@341 -- # ver2_l=1 00:06:35.366 05:34:08 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:35.366 05:34:08 thread -- scripts/common.sh@344 -- # case "$op" in 00:06:35.366 05:34:08 thread -- scripts/common.sh@345 -- # : 1 00:06:35.366 05:34:08 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:35.366 05:34:08 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:35.366 05:34:08 thread -- scripts/common.sh@365 -- # decimal 1 00:06:35.366 05:34:08 thread -- scripts/common.sh@353 -- # local d=1 00:06:35.366 05:34:08 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:35.366 05:34:08 thread -- scripts/common.sh@355 -- # echo 1 00:06:35.366 05:34:08 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:06:35.366 05:34:08 thread -- scripts/common.sh@366 -- # decimal 2 00:06:35.366 05:34:08 thread -- scripts/common.sh@353 -- # local d=2 00:06:35.366 05:34:08 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:35.366 05:34:08 thread -- scripts/common.sh@355 -- # echo 2 00:06:35.366 05:34:08 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:06:35.366 05:34:08 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:35.366 05:34:08 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:35.366 05:34:08 thread -- scripts/common.sh@368 -- # return 0 00:06:35.366 05:34:08 thread -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:35.366 05:34:08 thread -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:35.366 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:35.366 --rc genhtml_branch_coverage=1 00:06:35.366 --rc genhtml_function_coverage=1 00:06:35.366 --rc genhtml_legend=1 00:06:35.366 --rc geninfo_all_blocks=1 00:06:35.366 --rc geninfo_unexecuted_blocks=1 00:06:35.366 00:06:35.366 ' 00:06:35.366 05:34:08 thread -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:35.366 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:35.366 --rc genhtml_branch_coverage=1 00:06:35.366 --rc genhtml_function_coverage=1 00:06:35.366 --rc genhtml_legend=1 00:06:35.366 --rc geninfo_all_blocks=1 00:06:35.366 --rc geninfo_unexecuted_blocks=1 00:06:35.366 00:06:35.366 ' 00:06:35.366 05:34:08 thread -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:35.366 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:35.366 --rc genhtml_branch_coverage=1 00:06:35.366 --rc genhtml_function_coverage=1 00:06:35.366 --rc genhtml_legend=1 00:06:35.366 --rc geninfo_all_blocks=1 00:06:35.366 --rc geninfo_unexecuted_blocks=1 00:06:35.366 00:06:35.366 ' 00:06:35.366 05:34:08 thread -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:35.366 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:35.366 --rc genhtml_branch_coverage=1 00:06:35.366 --rc genhtml_function_coverage=1 00:06:35.366 --rc genhtml_legend=1 00:06:35.366 --rc geninfo_all_blocks=1 00:06:35.366 --rc geninfo_unexecuted_blocks=1 00:06:35.366 00:06:35.366 ' 00:06:35.366 05:34:08 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:35.366 05:34:08 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:06:35.366 05:34:08 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:35.366 05:34:08 thread -- common/autotest_common.sh@10 -- # set +x 00:06:35.625 ************************************ 00:06:35.625 START TEST thread_poller_perf 00:06:35.625 ************************************ 00:06:35.625 05:34:08 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:35.625 [2024-12-07 05:34:08.778288] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:35.625 [2024-12-07 05:34:08.778470] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71000 ] 00:06:35.625 [2024-12-07 05:34:08.931012] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:35.625 [2024-12-07 05:34:08.955891] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:35.625 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:06:37.007 [2024-12-07T05:34:10.375Z] ====================================== 00:06:37.007 [2024-12-07T05:34:10.375Z] busy:2297047500 (cyc) 00:06:37.007 [2024-12-07T05:34:10.375Z] total_run_count: 427000 00:06:37.007 [2024-12-07T05:34:10.375Z] tsc_hz: 2290000000 (cyc) 00:06:37.007 [2024-12-07T05:34:10.375Z] ====================================== 00:06:37.007 [2024-12-07T05:34:10.375Z] poller_cost: 5379 (cyc), 2348 (nsec) 00:06:37.007 00:06:37.007 real 0m1.284s 00:06:37.007 user 0m1.102s 00:06:37.007 sys 0m0.076s 00:06:37.007 05:34:10 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:37.007 05:34:10 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:06:37.007 ************************************ 00:06:37.007 END TEST thread_poller_perf 00:06:37.007 ************************************ 00:06:37.007 05:34:10 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:37.007 05:34:10 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:06:37.007 05:34:10 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:37.007 05:34:10 thread -- common/autotest_common.sh@10 -- # set +x 00:06:37.007 ************************************ 00:06:37.007 START TEST thread_poller_perf 00:06:37.007 ************************************ 00:06:37.007 05:34:10 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:37.007 [2024-12-07 05:34:10.126444] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:37.007 [2024-12-07 05:34:10.126599] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71042 ] 00:06:37.007 [2024-12-07 05:34:10.281906] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:37.007 [2024-12-07 05:34:10.306906] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:37.007 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:06:38.388 [2024-12-07T05:34:11.756Z] ====================================== 00:06:38.388 [2024-12-07T05:34:11.756Z] busy:2293549616 (cyc) 00:06:38.388 [2024-12-07T05:34:11.756Z] total_run_count: 5081000 00:06:38.388 [2024-12-07T05:34:11.756Z] tsc_hz: 2290000000 (cyc) 00:06:38.388 [2024-12-07T05:34:11.756Z] ====================================== 00:06:38.388 [2024-12-07T05:34:11.756Z] poller_cost: 451 (cyc), 196 (nsec) 00:06:38.388 00:06:38.388 real 0m1.282s 00:06:38.388 user 0m1.104s 00:06:38.388 sys 0m0.073s 00:06:38.388 05:34:11 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:38.388 ************************************ 00:06:38.388 END TEST thread_poller_perf 00:06:38.388 ************************************ 00:06:38.388 05:34:11 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:06:38.388 05:34:11 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:06:38.388 ************************************ 00:06:38.388 END TEST thread 00:06:38.388 ************************************ 00:06:38.388 00:06:38.388 real 0m2.920s 00:06:38.388 user 0m2.396s 00:06:38.388 sys 0m0.328s 00:06:38.388 05:34:11 thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:38.388 05:34:11 thread -- common/autotest_common.sh@10 -- # set +x 00:06:38.388 05:34:11 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:06:38.388 05:34:11 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:06:38.388 05:34:11 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:38.388 05:34:11 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:38.388 05:34:11 -- common/autotest_common.sh@10 -- # set +x 00:06:38.388 ************************************ 00:06:38.388 START TEST app_cmdline 00:06:38.388 ************************************ 00:06:38.388 05:34:11 app_cmdline -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:06:38.388 * Looking for test storage... 00:06:38.388 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:06:38.388 05:34:11 app_cmdline -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:38.388 05:34:11 app_cmdline -- common/autotest_common.sh@1711 -- # lcov --version 00:06:38.388 05:34:11 app_cmdline -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:38.388 05:34:11 app_cmdline -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@345 -- # : 1 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:06:38.388 05:34:11 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:38.389 05:34:11 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:06:38.389 05:34:11 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:06:38.389 05:34:11 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:38.389 05:34:11 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:38.389 05:34:11 app_cmdline -- scripts/common.sh@368 -- # return 0 00:06:38.389 05:34:11 app_cmdline -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:38.389 05:34:11 app_cmdline -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:38.389 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.389 --rc genhtml_branch_coverage=1 00:06:38.389 --rc genhtml_function_coverage=1 00:06:38.389 --rc genhtml_legend=1 00:06:38.389 --rc geninfo_all_blocks=1 00:06:38.389 --rc geninfo_unexecuted_blocks=1 00:06:38.389 00:06:38.389 ' 00:06:38.389 05:34:11 app_cmdline -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:38.389 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.389 --rc genhtml_branch_coverage=1 00:06:38.389 --rc genhtml_function_coverage=1 00:06:38.389 --rc genhtml_legend=1 00:06:38.389 --rc geninfo_all_blocks=1 00:06:38.389 --rc geninfo_unexecuted_blocks=1 00:06:38.389 00:06:38.389 ' 00:06:38.389 05:34:11 app_cmdline -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:38.389 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.389 --rc genhtml_branch_coverage=1 00:06:38.389 --rc genhtml_function_coverage=1 00:06:38.389 --rc genhtml_legend=1 00:06:38.389 --rc geninfo_all_blocks=1 00:06:38.389 --rc geninfo_unexecuted_blocks=1 00:06:38.389 00:06:38.389 ' 00:06:38.389 05:34:11 app_cmdline -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:38.389 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.389 --rc genhtml_branch_coverage=1 00:06:38.389 --rc genhtml_function_coverage=1 00:06:38.389 --rc genhtml_legend=1 00:06:38.389 --rc geninfo_all_blocks=1 00:06:38.389 --rc geninfo_unexecuted_blocks=1 00:06:38.389 00:06:38.389 ' 00:06:38.389 05:34:11 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:06:38.389 05:34:11 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=71120 00:06:38.389 05:34:11 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:06:38.389 05:34:11 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 71120 00:06:38.389 05:34:11 app_cmdline -- common/autotest_common.sh@835 -- # '[' -z 71120 ']' 00:06:38.389 05:34:11 app_cmdline -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:38.389 05:34:11 app_cmdline -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:38.389 05:34:11 app_cmdline -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:38.389 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:38.389 05:34:11 app_cmdline -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:38.389 05:34:11 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:38.648 [2024-12-07 05:34:11.805543] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:38.648 [2024-12-07 05:34:11.805705] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71120 ] 00:06:38.648 [2024-12-07 05:34:11.963530] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:38.648 [2024-12-07 05:34:11.988591] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@868 -- # return 0 00:06:39.586 05:34:12 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:06:39.586 { 00:06:39.586 "version": "SPDK v25.01-pre git sha1 a2f5e1c2d", 00:06:39.586 "fields": { 00:06:39.586 "major": 25, 00:06:39.586 "minor": 1, 00:06:39.586 "patch": 0, 00:06:39.586 "suffix": "-pre", 00:06:39.586 "commit": "a2f5e1c2d" 00:06:39.586 } 00:06:39.586 } 00:06:39.586 05:34:12 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:06:39.586 05:34:12 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:06:39.586 05:34:12 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:06:39.586 05:34:12 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:06:39.586 05:34:12 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:06:39.586 05:34:12 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.586 05:34:12 app_cmdline -- app/cmdline.sh@26 -- # sort 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.586 05:34:12 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:06:39.586 05:34:12 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:06:39.586 05:34:12 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@652 -- # local es=0 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:06:39.586 05:34:12 app_cmdline -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:39.846 request: 00:06:39.846 { 00:06:39.846 "method": "env_dpdk_get_mem_stats", 00:06:39.846 "req_id": 1 00:06:39.846 } 00:06:39.846 Got JSON-RPC error response 00:06:39.846 response: 00:06:39.846 { 00:06:39.846 "code": -32601, 00:06:39.846 "message": "Method not found" 00:06:39.846 } 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@655 -- # es=1 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:39.846 05:34:13 app_cmdline -- app/cmdline.sh@1 -- # killprocess 71120 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@954 -- # '[' -z 71120 ']' 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@958 -- # kill -0 71120 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@959 -- # uname 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71120 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71120' 00:06:39.846 killing process with pid 71120 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@973 -- # kill 71120 00:06:39.846 05:34:13 app_cmdline -- common/autotest_common.sh@978 -- # wait 71120 00:06:40.415 00:06:40.415 real 0m1.980s 00:06:40.415 user 0m2.220s 00:06:40.415 sys 0m0.545s 00:06:40.415 ************************************ 00:06:40.415 END TEST app_cmdline 00:06:40.415 ************************************ 00:06:40.415 05:34:13 app_cmdline -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:40.415 05:34:13 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:40.415 05:34:13 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:06:40.415 05:34:13 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:40.415 05:34:13 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:40.415 05:34:13 -- common/autotest_common.sh@10 -- # set +x 00:06:40.415 ************************************ 00:06:40.415 START TEST version 00:06:40.415 ************************************ 00:06:40.415 05:34:13 version -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:06:40.415 * Looking for test storage... 00:06:40.415 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:06:40.416 05:34:13 version -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:40.416 05:34:13 version -- common/autotest_common.sh@1711 -- # lcov --version 00:06:40.416 05:34:13 version -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:40.416 05:34:13 version -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:40.416 05:34:13 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:40.416 05:34:13 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:40.416 05:34:13 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:40.416 05:34:13 version -- scripts/common.sh@336 -- # IFS=.-: 00:06:40.416 05:34:13 version -- scripts/common.sh@336 -- # read -ra ver1 00:06:40.416 05:34:13 version -- scripts/common.sh@337 -- # IFS=.-: 00:06:40.416 05:34:13 version -- scripts/common.sh@337 -- # read -ra ver2 00:06:40.416 05:34:13 version -- scripts/common.sh@338 -- # local 'op=<' 00:06:40.416 05:34:13 version -- scripts/common.sh@340 -- # ver1_l=2 00:06:40.416 05:34:13 version -- scripts/common.sh@341 -- # ver2_l=1 00:06:40.416 05:34:13 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:40.416 05:34:13 version -- scripts/common.sh@344 -- # case "$op" in 00:06:40.416 05:34:13 version -- scripts/common.sh@345 -- # : 1 00:06:40.416 05:34:13 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:40.416 05:34:13 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:40.416 05:34:13 version -- scripts/common.sh@365 -- # decimal 1 00:06:40.416 05:34:13 version -- scripts/common.sh@353 -- # local d=1 00:06:40.416 05:34:13 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:40.416 05:34:13 version -- scripts/common.sh@355 -- # echo 1 00:06:40.416 05:34:13 version -- scripts/common.sh@365 -- # ver1[v]=1 00:06:40.416 05:34:13 version -- scripts/common.sh@366 -- # decimal 2 00:06:40.416 05:34:13 version -- scripts/common.sh@353 -- # local d=2 00:06:40.416 05:34:13 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:40.416 05:34:13 version -- scripts/common.sh@355 -- # echo 2 00:06:40.416 05:34:13 version -- scripts/common.sh@366 -- # ver2[v]=2 00:06:40.416 05:34:13 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:40.416 05:34:13 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:40.416 05:34:13 version -- scripts/common.sh@368 -- # return 0 00:06:40.416 05:34:13 version -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:40.416 05:34:13 version -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:40.416 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.416 --rc genhtml_branch_coverage=1 00:06:40.416 --rc genhtml_function_coverage=1 00:06:40.416 --rc genhtml_legend=1 00:06:40.416 --rc geninfo_all_blocks=1 00:06:40.416 --rc geninfo_unexecuted_blocks=1 00:06:40.416 00:06:40.416 ' 00:06:40.416 05:34:13 version -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:40.416 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.416 --rc genhtml_branch_coverage=1 00:06:40.416 --rc genhtml_function_coverage=1 00:06:40.416 --rc genhtml_legend=1 00:06:40.416 --rc geninfo_all_blocks=1 00:06:40.416 --rc geninfo_unexecuted_blocks=1 00:06:40.416 00:06:40.416 ' 00:06:40.416 05:34:13 version -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:40.416 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.416 --rc genhtml_branch_coverage=1 00:06:40.416 --rc genhtml_function_coverage=1 00:06:40.416 --rc genhtml_legend=1 00:06:40.416 --rc geninfo_all_blocks=1 00:06:40.416 --rc geninfo_unexecuted_blocks=1 00:06:40.416 00:06:40.416 ' 00:06:40.416 05:34:13 version -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:40.416 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.416 --rc genhtml_branch_coverage=1 00:06:40.416 --rc genhtml_function_coverage=1 00:06:40.416 --rc genhtml_legend=1 00:06:40.416 --rc geninfo_all_blocks=1 00:06:40.416 --rc geninfo_unexecuted_blocks=1 00:06:40.416 00:06:40.416 ' 00:06:40.416 05:34:13 version -- app/version.sh@17 -- # get_header_version major 00:06:40.416 05:34:13 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:40.416 05:34:13 version -- app/version.sh@14 -- # cut -f2 00:06:40.416 05:34:13 version -- app/version.sh@14 -- # tr -d '"' 00:06:40.416 05:34:13 version -- app/version.sh@17 -- # major=25 00:06:40.416 05:34:13 version -- app/version.sh@18 -- # get_header_version minor 00:06:40.416 05:34:13 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:40.416 05:34:13 version -- app/version.sh@14 -- # cut -f2 00:06:40.416 05:34:13 version -- app/version.sh@14 -- # tr -d '"' 00:06:40.416 05:34:13 version -- app/version.sh@18 -- # minor=1 00:06:40.676 05:34:13 version -- app/version.sh@19 -- # get_header_version patch 00:06:40.676 05:34:13 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:40.676 05:34:13 version -- app/version.sh@14 -- # cut -f2 00:06:40.676 05:34:13 version -- app/version.sh@14 -- # tr -d '"' 00:06:40.676 05:34:13 version -- app/version.sh@19 -- # patch=0 00:06:40.676 05:34:13 version -- app/version.sh@20 -- # get_header_version suffix 00:06:40.676 05:34:13 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:40.676 05:34:13 version -- app/version.sh@14 -- # cut -f2 00:06:40.676 05:34:13 version -- app/version.sh@14 -- # tr -d '"' 00:06:40.676 05:34:13 version -- app/version.sh@20 -- # suffix=-pre 00:06:40.676 05:34:13 version -- app/version.sh@22 -- # version=25.1 00:06:40.676 05:34:13 version -- app/version.sh@25 -- # (( patch != 0 )) 00:06:40.676 05:34:13 version -- app/version.sh@28 -- # version=25.1rc0 00:06:40.676 05:34:13 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:06:40.676 05:34:13 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:06:40.676 05:34:13 version -- app/version.sh@30 -- # py_version=25.1rc0 00:06:40.676 05:34:13 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:06:40.676 ************************************ 00:06:40.676 END TEST version 00:06:40.676 ************************************ 00:06:40.676 00:06:40.676 real 0m0.308s 00:06:40.676 user 0m0.192s 00:06:40.676 sys 0m0.170s 00:06:40.676 05:34:13 version -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:40.676 05:34:13 version -- common/autotest_common.sh@10 -- # set +x 00:06:40.676 05:34:13 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:06:40.676 05:34:13 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:06:40.676 05:34:13 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:06:40.676 05:34:13 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:40.676 05:34:13 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:40.676 05:34:13 -- common/autotest_common.sh@10 -- # set +x 00:06:40.676 ************************************ 00:06:40.676 START TEST bdev_raid 00:06:40.676 ************************************ 00:06:40.676 05:34:13 bdev_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:06:40.676 * Looking for test storage... 00:06:40.676 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:06:40.676 05:34:14 bdev_raid -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:40.676 05:34:14 bdev_raid -- common/autotest_common.sh@1711 -- # lcov --version 00:06:40.676 05:34:14 bdev_raid -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:40.936 05:34:14 bdev_raid -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@345 -- # : 1 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:40.936 05:34:14 bdev_raid -- scripts/common.sh@368 -- # return 0 00:06:40.936 05:34:14 bdev_raid -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:40.936 05:34:14 bdev_raid -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:40.936 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.936 --rc genhtml_branch_coverage=1 00:06:40.936 --rc genhtml_function_coverage=1 00:06:40.936 --rc genhtml_legend=1 00:06:40.936 --rc geninfo_all_blocks=1 00:06:40.936 --rc geninfo_unexecuted_blocks=1 00:06:40.936 00:06:40.936 ' 00:06:40.936 05:34:14 bdev_raid -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:40.936 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.936 --rc genhtml_branch_coverage=1 00:06:40.936 --rc genhtml_function_coverage=1 00:06:40.936 --rc genhtml_legend=1 00:06:40.936 --rc geninfo_all_blocks=1 00:06:40.936 --rc geninfo_unexecuted_blocks=1 00:06:40.936 00:06:40.936 ' 00:06:40.936 05:34:14 bdev_raid -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:40.936 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.936 --rc genhtml_branch_coverage=1 00:06:40.936 --rc genhtml_function_coverage=1 00:06:40.936 --rc genhtml_legend=1 00:06:40.937 --rc geninfo_all_blocks=1 00:06:40.937 --rc geninfo_unexecuted_blocks=1 00:06:40.937 00:06:40.937 ' 00:06:40.937 05:34:14 bdev_raid -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:40.937 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:40.937 --rc genhtml_branch_coverage=1 00:06:40.937 --rc genhtml_function_coverage=1 00:06:40.937 --rc genhtml_legend=1 00:06:40.937 --rc geninfo_all_blocks=1 00:06:40.937 --rc geninfo_unexecuted_blocks=1 00:06:40.937 00:06:40.937 ' 00:06:40.937 05:34:14 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:06:40.937 05:34:14 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:06:40.937 05:34:14 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:06:40.937 05:34:14 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:06:40.937 05:34:14 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:06:40.937 05:34:14 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:06:40.937 05:34:14 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:06:40.937 05:34:14 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:40.937 05:34:14 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:40.937 05:34:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:40.937 ************************************ 00:06:40.937 START TEST raid1_resize_data_offset_test 00:06:40.937 ************************************ 00:06:40.937 05:34:14 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1129 -- # raid_resize_data_offset_test 00:06:40.937 05:34:14 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=71286 00:06:40.937 05:34:14 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:40.937 Process raid pid: 71286 00:06:40.937 05:34:14 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 71286' 00:06:40.937 05:34:14 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 71286 00:06:40.937 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:40.937 05:34:14 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@835 -- # '[' -z 71286 ']' 00:06:40.937 05:34:14 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:40.937 05:34:14 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:40.937 05:34:14 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:40.937 05:34:14 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:40.937 05:34:14 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:40.937 [2024-12-07 05:34:14.221520] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:40.937 [2024-12-07 05:34:14.221673] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:41.197 [2024-12-07 05:34:14.359966] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:41.197 [2024-12-07 05:34:14.384782] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:41.197 [2024-12-07 05:34:14.426535] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:41.197 [2024-12-07 05:34:14.426570] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@868 -- # return 0 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:41.767 malloc0 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:41.767 malloc1 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:41.767 null0 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:41.767 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:41.767 [2024-12-07 05:34:15.130830] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:06:41.767 [2024-12-07 05:34:15.132707] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:06:41.767 [2024-12-07 05:34:15.132748] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:06:41.767 [2024-12-07 05:34:15.132900] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:41.767 [2024-12-07 05:34:15.132934] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:06:42.027 [2024-12-07 05:34:15.133189] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:06:42.027 [2024-12-07 05:34:15.133321] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:42.027 [2024-12-07 05:34:15.133334] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:42.027 [2024-12-07 05:34:15.133454] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:42.027 [2024-12-07 05:34:15.194720] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:42.027 malloc2 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:42.027 [2024-12-07 05:34:15.317775] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:06:42.027 [2024-12-07 05:34:15.322922] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:42.027 [2024-12-07 05:34:15.324865] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 71286 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@954 -- # '[' -z 71286 ']' 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@958 -- # kill -0 71286 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # uname 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:42.027 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71286 00:06:42.286 killing process with pid 71286 00:06:42.286 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:42.286 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:42.286 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71286' 00:06:42.286 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@973 -- # kill 71286 00:06:42.286 [2024-12-07 05:34:15.413897] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:42.286 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@978 -- # wait 71286 00:06:42.286 [2024-12-07 05:34:15.414567] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:06:42.286 [2024-12-07 05:34:15.414641] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:42.286 [2024-12-07 05:34:15.414660] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:06:42.286 [2024-12-07 05:34:15.420498] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:42.286 [2024-12-07 05:34:15.420833] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:42.286 [2024-12-07 05:34:15.420854] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:42.286 [2024-12-07 05:34:15.629675] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:42.545 05:34:15 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:06:42.545 00:06:42.545 real 0m1.697s 00:06:42.545 user 0m1.713s 00:06:42.545 sys 0m0.412s 00:06:42.545 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:42.545 05:34:15 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:42.545 ************************************ 00:06:42.545 END TEST raid1_resize_data_offset_test 00:06:42.546 ************************************ 00:06:42.546 05:34:15 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:06:42.546 05:34:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:42.546 05:34:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:42.546 05:34:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:42.546 ************************************ 00:06:42.546 START TEST raid0_resize_superblock_test 00:06:42.546 ************************************ 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 0 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71336 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71336' 00:06:42.546 Process raid pid: 71336 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71336 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 71336 ']' 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:42.546 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:42.546 05:34:15 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:42.805 [2024-12-07 05:34:15.984321] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:42.805 [2024-12-07 05:34:15.984516] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:42.805 [2024-12-07 05:34:16.140183] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:42.805 [2024-12-07 05:34:16.165943] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:43.064 [2024-12-07 05:34:16.208009] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:43.064 [2024-12-07 05:34:16.208123] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:43.633 05:34:16 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:43.633 05:34:16 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:06:43.633 05:34:16 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:06:43.633 05:34:16 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.633 05:34:16 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.633 malloc0 00:06:43.633 05:34:16 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.634 05:34:16 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:43.634 05:34:16 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.634 05:34:16 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.634 [2024-12-07 05:34:16.932658] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:43.634 [2024-12-07 05:34:16.932716] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:43.634 [2024-12-07 05:34:16.932753] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:06:43.634 [2024-12-07 05:34:16.932763] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:43.634 [2024-12-07 05:34:16.934867] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:43.634 [2024-12-07 05:34:16.934908] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:43.634 pt0 00:06:43.634 05:34:16 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.634 05:34:16 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:06:43.634 05:34:16 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.634 05:34:16 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.892 c3d0c394-741a-4db6-b767-949da76b75a8 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.892 947a0992-c3d8-4f64-9e91-b9bff76a5fe3 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.892 40be5633-0e84-4967-9310-f16e546189a6 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.892 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.892 [2024-12-07 05:34:17.067000] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 947a0992-c3d8-4f64-9e91-b9bff76a5fe3 is claimed 00:06:43.892 [2024-12-07 05:34:17.067077] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 40be5633-0e84-4967-9310-f16e546189a6 is claimed 00:06:43.892 [2024-12-07 05:34:17.067204] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:43.892 [2024-12-07 05:34:17.067231] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:06:43.892 [2024-12-07 05:34:17.067472] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:43.893 [2024-12-07 05:34:17.067612] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:43.893 [2024-12-07 05:34:17.067622] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:43.893 [2024-12-07 05:34:17.067784] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:06:43.893 [2024-12-07 05:34:17.175058] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.893 [2024-12-07 05:34:17.222862] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:43.893 [2024-12-07 05:34:17.222886] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '947a0992-c3d8-4f64-9e91-b9bff76a5fe3' was resized: old size 131072, new size 204800 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:43.893 [2024-12-07 05:34:17.230809] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:43.893 [2024-12-07 05:34:17.230830] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '40be5633-0e84-4967-9310-f16e546189a6' was resized: old size 131072, new size 204800 00:06:43.893 [2024-12-07 05:34:17.230858] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:43.893 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.152 [2024-12-07 05:34:17.346725] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.152 [2024-12-07 05:34:17.394424] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:06:44.152 [2024-12-07 05:34:17.394496] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:06:44.152 [2024-12-07 05:34:17.394507] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:44.152 [2024-12-07 05:34:17.394524] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:06:44.152 [2024-12-07 05:34:17.394638] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:44.152 [2024-12-07 05:34:17.394743] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:44.152 [2024-12-07 05:34:17.394761] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.152 [2024-12-07 05:34:17.406361] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:44.152 [2024-12-07 05:34:17.406414] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:44.152 [2024-12-07 05:34:17.406433] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:06:44.152 [2024-12-07 05:34:17.406442] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:44.152 [2024-12-07 05:34:17.408568] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:44.152 [2024-12-07 05:34:17.408604] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:44.152 [2024-12-07 05:34:17.409975] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 947a0992-c3d8-4f64-9e91-b9bff76a5fe3 00:06:44.152 [2024-12-07 05:34:17.410038] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 947a0992-c3d8-4f64-9e91-b9bff76a5fe3 is claimed 00:06:44.152 [2024-12-07 05:34:17.410131] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 40be5633-0e84-4967-9310-f16e546189a6 00:06:44.152 [2024-12-07 05:34:17.410156] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 40be5633-0e84-4967-9310-f16e546189a6 is claimed 00:06:44.152 [2024-12-07 05:34:17.410245] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 40be5633-0e84-4967-9310-f16e546189a6 (2) smaller than existing raid bdev Raid (3) 00:06:44.152 [2024-12-07 05:34:17.410263] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 947a0992-c3d8-4f64-9e91-b9bff76a5fe3: File exists 00:06:44.152 [2024-12-07 05:34:17.410298] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:06:44.152 [2024-12-07 05:34:17.410323] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:06:44.152 [2024-12-07 05:34:17.410524] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:06:44.152 [2024-12-07 05:34:17.410683] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:06:44.152 [2024-12-07 05:34:17.410698] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:06:44.152 [2024-12-07 05:34:17.410839] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:44.152 pt0 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.152 [2024-12-07 05:34:17.434979] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71336 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 71336 ']' 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 71336 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71336 00:06:44.152 killing process with pid 71336 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71336' 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 71336 00:06:44.152 [2024-12-07 05:34:17.513574] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:44.152 [2024-12-07 05:34:17.513675] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:44.152 [2024-12-07 05:34:17.513725] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:44.152 [2024-12-07 05:34:17.513735] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state offline 00:06:44.152 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 71336 00:06:44.412 [2024-12-07 05:34:17.671378] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:44.704 05:34:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:06:44.704 00:06:44.704 real 0m1.977s 00:06:44.704 user 0m2.254s 00:06:44.704 sys 0m0.486s 00:06:44.704 ************************************ 00:06:44.704 END TEST raid0_resize_superblock_test 00:06:44.704 ************************************ 00:06:44.704 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:44.704 05:34:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.704 05:34:17 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:06:44.704 05:34:17 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:44.704 05:34:17 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:44.704 05:34:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:44.704 ************************************ 00:06:44.704 START TEST raid1_resize_superblock_test 00:06:44.704 ************************************ 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 1 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71407 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:44.704 Process raid pid: 71407 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71407' 00:06:44.704 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71407 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 71407 ']' 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:44.704 05:34:17 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.704 [2024-12-07 05:34:18.032356] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:44.704 [2024-12-07 05:34:18.032578] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:44.970 [2024-12-07 05:34:18.190377] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:44.970 [2024-12-07 05:34:18.215639] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:44.970 [2024-12-07 05:34:18.257218] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:44.970 [2024-12-07 05:34:18.257273] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:45.539 05:34:18 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:45.539 05:34:18 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:06:45.539 05:34:18 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:06:45.539 05:34:18 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.539 05:34:18 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.796 malloc0 00:06:45.796 05:34:18 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:45.796 05:34:18 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:45.796 05:34:18 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.796 05:34:18 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.796 [2024-12-07 05:34:18.975730] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:45.796 [2024-12-07 05:34:18.975850] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:45.796 [2024-12-07 05:34:18.975894] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:06:45.797 [2024-12-07 05:34:18.975908] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:45.797 [2024-12-07 05:34:18.978205] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:45.797 [2024-12-07 05:34:18.978246] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:45.797 pt0 00:06:45.797 05:34:18 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:45.797 05:34:18 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:06:45.797 05:34:18 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.797 05:34:18 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.797 ff840f4a-2da6-4cfc-9467-668930025b2c 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.797 ffbdc362-911e-403c-943f-4c516ffa0c90 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.797 24344515-19f1-48a6-bc05-895301aa8012 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.797 [2024-12-07 05:34:19.095465] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev ffbdc362-911e-403c-943f-4c516ffa0c90 is claimed 00:06:45.797 [2024-12-07 05:34:19.095553] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 24344515-19f1-48a6-bc05-895301aa8012 is claimed 00:06:45.797 [2024-12-07 05:34:19.095696] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:45.797 [2024-12-07 05:34:19.095710] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:06:45.797 [2024-12-07 05:34:19.095983] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:45.797 [2024-12-07 05:34:19.096149] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:45.797 [2024-12-07 05:34:19.096158] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:45.797 [2024-12-07 05:34:19.096300] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.797 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.056 [2024-12-07 05:34:19.207551] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.056 [2024-12-07 05:34:19.251323] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:46.056 [2024-12-07 05:34:19.251392] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'ffbdc362-911e-403c-943f-4c516ffa0c90' was resized: old size 131072, new size 204800 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.056 [2024-12-07 05:34:19.259272] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:46.056 [2024-12-07 05:34:19.259294] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '24344515-19f1-48a6-bc05-895301aa8012' was resized: old size 131072, new size 204800 00:06:46.056 [2024-12-07 05:34:19.259321] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.056 [2024-12-07 05:34:19.347220] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.056 [2024-12-07 05:34:19.390965] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:06:46.056 [2024-12-07 05:34:19.391030] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:06:46.056 [2024-12-07 05:34:19.391053] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:06:46.056 [2024-12-07 05:34:19.391209] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:46.056 [2024-12-07 05:34:19.391346] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:46.056 [2024-12-07 05:34:19.391397] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:46.056 [2024-12-07 05:34:19.391409] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.056 [2024-12-07 05:34:19.402897] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:46.056 [2024-12-07 05:34:19.402945] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:46.056 [2024-12-07 05:34:19.402962] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:06:46.056 [2024-12-07 05:34:19.402972] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:46.056 [2024-12-07 05:34:19.405129] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:46.056 [2024-12-07 05:34:19.405167] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:46.056 [2024-12-07 05:34:19.406575] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev ffbdc362-911e-403c-943f-4c516ffa0c90 00:06:46.056 [2024-12-07 05:34:19.406698] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev ffbdc362-911e-403c-943f-4c516ffa0c90 is claimed 00:06:46.056 [2024-12-07 05:34:19.406784] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 24344515-19f1-48a6-bc05-895301aa8012 00:06:46.056 [2024-12-07 05:34:19.406815] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 24344515-19f1-48a6-bc05-895301aa8012 is claimed 00:06:46.056 [2024-12-07 05:34:19.406907] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 24344515-19f1-48a6-bc05-895301aa8012 (2) smaller than existing raid bdev Raid (3) 00:06:46.056 [2024-12-07 05:34:19.406925] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev ffbdc362-911e-403c-943f-4c516ffa0c90: File exists 00:06:46.056 [2024-12-07 05:34:19.406968] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:06:46.056 [2024-12-07 05:34:19.406977] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:06:46.056 [2024-12-07 05:34:19.407192] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:06:46.056 [2024-12-07 05:34:19.407345] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:06:46.056 [2024-12-07 05:34:19.407361] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:06:46.056 [2024-12-07 05:34:19.407508] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:46.056 pt0 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.056 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:06:46.316 [2024-12-07 05:34:19.427170] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71407 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 71407 ']' 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 71407 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71407 00:06:46.316 killing process with pid 71407 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71407' 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 71407 00:06:46.316 [2024-12-07 05:34:19.509284] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:46.316 [2024-12-07 05:34:19.509357] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:46.316 [2024-12-07 05:34:19.509405] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:46.316 [2024-12-07 05:34:19.509413] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state offline 00:06:46.316 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 71407 00:06:46.316 [2024-12-07 05:34:19.668251] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:46.575 05:34:19 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:06:46.575 00:06:46.575 real 0m1.928s 00:06:46.575 user 0m2.195s 00:06:46.575 sys 0m0.472s 00:06:46.575 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:46.575 ************************************ 00:06:46.575 END TEST raid1_resize_superblock_test 00:06:46.575 ************************************ 00:06:46.575 05:34:19 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.575 05:34:19 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:06:46.834 05:34:19 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:06:46.834 05:34:19 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:06:46.834 05:34:19 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:06:46.834 05:34:19 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:06:46.834 05:34:19 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:06:46.834 05:34:19 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:46.834 05:34:19 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:46.834 05:34:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:46.834 ************************************ 00:06:46.834 START TEST raid_function_test_raid0 00:06:46.834 ************************************ 00:06:46.834 05:34:19 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1129 -- # raid_function_test raid0 00:06:46.834 05:34:19 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:06:46.834 05:34:19 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:06:46.834 05:34:19 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:06:46.834 05:34:19 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=71483 00:06:46.834 Process raid pid: 71483 00:06:46.835 05:34:19 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:46.835 05:34:19 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71483' 00:06:46.835 05:34:19 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 71483 00:06:46.835 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:46.835 05:34:19 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@835 -- # '[' -z 71483 ']' 00:06:46.835 05:34:19 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:46.835 05:34:19 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:46.835 05:34:19 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:46.835 05:34:19 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:46.835 05:34:19 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:46.835 [2024-12-07 05:34:20.058297] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:46.835 [2024-12-07 05:34:20.058507] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:47.110 [2024-12-07 05:34:20.212912] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:47.110 [2024-12-07 05:34:20.237879] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:47.110 [2024-12-07 05:34:20.279864] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:47.110 [2024-12-07 05:34:20.279900] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@868 -- # return 0 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:47.707 Base_1 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:47.707 Base_2 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:47.707 [2024-12-07 05:34:20.915464] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:47.707 [2024-12-07 05:34:20.917344] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:47.707 [2024-12-07 05:34:20.917415] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:47.707 [2024-12-07 05:34:20.917426] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:47.707 [2024-12-07 05:34:20.917668] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:47.707 [2024-12-07 05:34:20.917791] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:47.707 [2024-12-07 05:34:20.917800] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:06:47.707 [2024-12-07 05:34:20.917932] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:47.707 05:34:20 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:06:47.967 [2024-12-07 05:34:21.151098] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:06:47.967 /dev/nbd0 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # local i 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@877 -- # break 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:06:47.967 1+0 records in 00:06:47.967 1+0 records out 00:06:47.967 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000354647 s, 11.5 MB/s 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # size=4096 00:06:47.967 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:47.968 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:47.968 05:34:21 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@893 -- # return 0 00:06:47.968 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:47.968 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:47.968 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:06:47.968 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:47.968 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:48.227 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:48.227 { 00:06:48.227 "nbd_device": "/dev/nbd0", 00:06:48.227 "bdev_name": "raid" 00:06:48.227 } 00:06:48.227 ]' 00:06:48.227 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:48.227 { 00:06:48.227 "nbd_device": "/dev/nbd0", 00:06:48.227 "bdev_name": "raid" 00:06:48.227 } 00:06:48.227 ]' 00:06:48.227 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:48.227 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:06:48.227 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:48.227 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:06:48.227 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:06:48.227 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:06:48.228 4096+0 records in 00:06:48.228 4096+0 records out 00:06:48.228 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0334883 s, 62.6 MB/s 00:06:48.228 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:06:48.488 4096+0 records in 00:06:48.488 4096+0 records out 00:06:48.488 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.178508 s, 11.7 MB/s 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:06:48.488 128+0 records in 00:06:48.488 128+0 records out 00:06:48.488 65536 bytes (66 kB, 64 KiB) copied, 0.00127345 s, 51.5 MB/s 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:06:48.488 2035+0 records in 00:06:48.488 2035+0 records out 00:06:48.488 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0140286 s, 74.3 MB/s 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:06:48.488 456+0 records in 00:06:48.488 456+0 records out 00:06:48.488 233472 bytes (233 kB, 228 KiB) copied, 0.00375012 s, 62.3 MB/s 00:06:48.488 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:48.489 05:34:21 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:06:48.750 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:48.750 [2024-12-07 05:34:22.036502] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:48.750 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:48.750 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:48.750 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:48.750 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:48.750 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:48.750 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:06:48.750 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:06:48.750 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:06:48.750 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:48.750 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 71483 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # '[' -z 71483 ']' 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@958 -- # kill -0 71483 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # uname 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71483 00:06:49.009 killing process with pid 71483 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71483' 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@973 -- # kill 71483 00:06:49.009 [2024-12-07 05:34:22.331642] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:49.009 [2024-12-07 05:34:22.331754] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:49.009 [2024-12-07 05:34:22.331808] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:49.009 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@978 -- # wait 71483 00:06:49.009 [2024-12-07 05:34:22.331820] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:06:49.009 [2024-12-07 05:34:22.354177] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:49.271 05:34:22 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:06:49.271 00:06:49.271 real 0m2.589s 00:06:49.271 user 0m3.234s 00:06:49.271 sys 0m0.863s 00:06:49.271 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:49.271 ************************************ 00:06:49.271 END TEST raid_function_test_raid0 00:06:49.271 ************************************ 00:06:49.271 05:34:22 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:49.271 05:34:22 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:06:49.271 05:34:22 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:49.271 05:34:22 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:49.271 05:34:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:49.271 ************************************ 00:06:49.271 START TEST raid_function_test_concat 00:06:49.271 ************************************ 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1129 -- # raid_function_test concat 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=71600 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71600' 00:06:49.271 Process raid pid: 71600 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 71600 00:06:49.271 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@835 -- # '[' -z 71600 ']' 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:49.271 05:34:22 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:49.532 [2024-12-07 05:34:22.708173] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:49.532 [2024-12-07 05:34:22.708373] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:49.532 [2024-12-07 05:34:22.843481] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:49.532 [2024-12-07 05:34:22.868374] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:49.792 [2024-12-07 05:34:22.910115] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:49.792 [2024-12-07 05:34:22.910252] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@868 -- # return 0 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:50.361 Base_1 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:50.361 Base_2 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:50.361 [2024-12-07 05:34:23.566179] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:50.361 [2024-12-07 05:34:23.567954] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:50.361 [2024-12-07 05:34:23.568018] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:50.361 [2024-12-07 05:34:23.568036] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:50.361 [2024-12-07 05:34:23.568291] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:50.361 [2024-12-07 05:34:23.568425] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:50.361 [2024-12-07 05:34:23.568434] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:06:50.361 [2024-12-07 05:34:23.568553] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:50.361 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:50.362 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:06:50.622 [2024-12-07 05:34:23.801823] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:06:50.622 /dev/nbd0 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # local i 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@877 -- # break 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:06:50.622 1+0 records in 00:06:50.622 1+0 records out 00:06:50.622 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000301314 s, 13.6 MB/s 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # size=4096 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@893 -- # return 0 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:50.622 05:34:23 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:50.882 { 00:06:50.882 "nbd_device": "/dev/nbd0", 00:06:50.882 "bdev_name": "raid" 00:06:50.882 } 00:06:50.882 ]' 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:50.882 { 00:06:50.882 "nbd_device": "/dev/nbd0", 00:06:50.882 "bdev_name": "raid" 00:06:50.882 } 00:06:50.882 ]' 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:06:50.882 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:06:50.882 4096+0 records in 00:06:50.882 4096+0 records out 00:06:50.882 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0351678 s, 59.6 MB/s 00:06:50.883 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:06:51.141 4096+0 records in 00:06:51.141 4096+0 records out 00:06:51.141 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.1815 s, 11.6 MB/s 00:06:51.141 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:06:51.141 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:51.141 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:06:51.142 128+0 records in 00:06:51.142 128+0 records out 00:06:51.142 65536 bytes (66 kB, 64 KiB) copied, 0.00109902 s, 59.6 MB/s 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:06:51.142 2035+0 records in 00:06:51.142 2035+0 records out 00:06:51.142 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0122272 s, 85.2 MB/s 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:06:51.142 456+0 records in 00:06:51.142 456+0 records out 00:06:51.142 233472 bytes (233 kB, 228 KiB) copied, 0.00361305 s, 64.6 MB/s 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:51.142 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:06:51.400 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:51.400 [2024-12-07 05:34:24.663052] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:51.400 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:51.400 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:51.400 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:51.400 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:51.400 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:51.400 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:06:51.400 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:06:51.400 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:06:51.400 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:51.400 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 71600 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # '[' -z 71600 ']' 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@958 -- # kill -0 71600 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # uname 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71600 00:06:51.659 killing process with pid 71600 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71600' 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@973 -- # kill 71600 00:06:51.659 [2024-12-07 05:34:24.970201] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:51.659 [2024-12-07 05:34:24.970325] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:51.659 [2024-12-07 05:34:24.970379] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:51.659 [2024-12-07 05:34:24.970391] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:06:51.659 05:34:24 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@978 -- # wait 71600 00:06:51.659 [2024-12-07 05:34:24.992413] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:51.916 05:34:25 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:06:51.916 00:06:51.916 real 0m2.577s 00:06:51.916 user 0m3.228s 00:06:51.916 sys 0m0.839s 00:06:51.916 ************************************ 00:06:51.916 END TEST raid_function_test_concat 00:06:51.916 ************************************ 00:06:51.916 05:34:25 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:51.916 05:34:25 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:51.916 05:34:25 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:06:51.916 05:34:25 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:51.916 05:34:25 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:51.916 05:34:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:51.916 ************************************ 00:06:51.916 START TEST raid0_resize_test 00:06:51.916 ************************************ 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 0 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=71711 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 71711' 00:06:51.916 Process raid pid: 71711 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 71711 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- common/autotest_common.sh@835 -- # '[' -z 71711 ']' 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:51.916 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:51.916 05:34:25 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:52.175 [2024-12-07 05:34:25.328448] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:52.175 [2024-12-07 05:34:25.328653] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:52.175 [2024-12-07 05:34:25.482848] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:52.175 [2024-12-07 05:34:25.507333] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:52.434 [2024-12-07 05:34:25.548614] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:52.434 [2024-12-07 05:34:25.548735] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@868 -- # return 0 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.002 Base_1 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.002 Base_2 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.002 [2024-12-07 05:34:26.194489] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:53.002 [2024-12-07 05:34:26.196284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:53.002 [2024-12-07 05:34:26.196335] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:53.002 [2024-12-07 05:34:26.196358] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:53.002 [2024-12-07 05:34:26.196592] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:06:53.002 [2024-12-07 05:34:26.196693] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:53.002 [2024-12-07 05:34:26.196701] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:53.002 [2024-12-07 05:34:26.196816] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.002 [2024-12-07 05:34:26.206459] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:53.002 [2024-12-07 05:34:26.206482] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:06:53.002 true 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.002 [2024-12-07 05:34:26.222632] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.002 [2024-12-07 05:34:26.266336] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:53.002 [2024-12-07 05:34:26.266357] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:06:53.002 [2024-12-07 05:34:26.266383] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:06:53.002 true 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:06:53.002 [2024-12-07 05:34:26.278490] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 71711 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # '[' -z 71711 ']' 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@958 -- # kill -0 71711 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # uname 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71711 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71711' 00:06:53.002 killing process with pid 71711 00:06:53.002 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@973 -- # kill 71711 00:06:53.002 [2024-12-07 05:34:26.366712] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:53.003 [2024-12-07 05:34:26.366846] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:53.003 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@978 -- # wait 71711 00:06:53.003 [2024-12-07 05:34:26.366928] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:53.003 [2024-12-07 05:34:26.366946] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:53.262 [2024-12-07 05:34:26.368424] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:53.262 05:34:26 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:06:53.262 00:06:53.262 real 0m1.315s 00:06:53.262 user 0m1.492s 00:06:53.262 sys 0m0.278s 00:06:53.262 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:53.262 05:34:26 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.262 ************************************ 00:06:53.262 END TEST raid0_resize_test 00:06:53.262 ************************************ 00:06:53.523 05:34:26 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:06:53.523 05:34:26 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:06:53.523 05:34:26 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:53.523 05:34:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:53.523 ************************************ 00:06:53.523 START TEST raid1_resize_test 00:06:53.523 ************************************ 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 1 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:06:53.523 Process raid pid: 71762 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=71762 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 71762' 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 71762 00:06:53.523 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- common/autotest_common.sh@835 -- # '[' -z 71762 ']' 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:53.523 05:34:26 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:53.523 [2024-12-07 05:34:26.727925] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:53.523 [2024-12-07 05:34:26.728149] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:53.523 [2024-12-07 05:34:26.876822] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:53.784 [2024-12-07 05:34:26.901381] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:53.784 [2024-12-07 05:34:26.943706] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:53.784 [2024-12-07 05:34:26.943800] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@868 -- # return 0 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:54.355 Base_1 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:54.355 Base_2 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:54.355 [2024-12-07 05:34:27.586278] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:54.355 [2024-12-07 05:34:27.588158] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:54.355 [2024-12-07 05:34:27.588214] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:54.355 [2024-12-07 05:34:27.588226] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:06:54.355 [2024-12-07 05:34:27.588488] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:06:54.355 [2024-12-07 05:34:27.588602] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:54.355 [2024-12-07 05:34:27.588612] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:54.355 [2024-12-07 05:34:27.588731] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:54.355 [2024-12-07 05:34:27.598278] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:54.355 [2024-12-07 05:34:27.598309] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:06:54.355 true 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:54.355 [2024-12-07 05:34:27.610412] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:54.355 [2024-12-07 05:34:27.658124] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:54.355 [2024-12-07 05:34:27.658186] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:06:54.355 [2024-12-07 05:34:27.658248] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:06:54.355 true 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:54.355 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:54.355 [2024-12-07 05:34:27.670289] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:54.356 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:54.356 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:06:54.356 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:06:54.356 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:06:54.356 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:06:54.356 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:06:54.356 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 71762 00:06:54.356 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@954 -- # '[' -z 71762 ']' 00:06:54.356 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@958 -- # kill -0 71762 00:06:54.356 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # uname 00:06:54.356 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:54.615 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71762 00:06:54.615 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:54.615 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:54.615 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71762' 00:06:54.615 killing process with pid 71762 00:06:54.615 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@973 -- # kill 71762 00:06:54.615 [2024-12-07 05:34:27.756811] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:54.615 [2024-12-07 05:34:27.756940] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:54.615 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@978 -- # wait 71762 00:06:54.615 [2024-12-07 05:34:27.757381] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:54.615 [2024-12-07 05:34:27.757451] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:54.615 [2024-12-07 05:34:27.758583] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:54.615 05:34:27 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:06:54.615 00:06:54.615 real 0m1.323s 00:06:54.615 user 0m1.518s 00:06:54.615 sys 0m0.274s 00:06:54.615 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:54.615 05:34:27 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:54.615 ************************************ 00:06:54.615 END TEST raid1_resize_test 00:06:54.615 ************************************ 00:06:54.875 05:34:28 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:06:54.875 05:34:28 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:06:54.875 05:34:28 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:06:54.875 05:34:28 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:54.875 05:34:28 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:54.875 05:34:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:54.875 ************************************ 00:06:54.875 START TEST raid_state_function_test 00:06:54.875 ************************************ 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 false 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:06:54.875 Process raid pid: 71813 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=71813 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 71813' 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 71813 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 71813 ']' 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:54.875 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:54.875 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:54.875 [2024-12-07 05:34:28.121657] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:54.875 [2024-12-07 05:34:28.122220] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:55.135 [2024-12-07 05:34:28.279127] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:55.135 [2024-12-07 05:34:28.304271] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:55.135 [2024-12-07 05:34:28.345969] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:55.135 [2024-12-07 05:34:28.346003] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:55.719 [2024-12-07 05:34:28.988012] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:55.719 [2024-12-07 05:34:28.988138] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:55.719 [2024-12-07 05:34:28.988169] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:55.719 [2024-12-07 05:34:28.988194] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:55.719 05:34:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:55.719 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:55.719 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:55.719 "name": "Existed_Raid", 00:06:55.719 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:55.719 "strip_size_kb": 64, 00:06:55.719 "state": "configuring", 00:06:55.719 "raid_level": "raid0", 00:06:55.719 "superblock": false, 00:06:55.719 "num_base_bdevs": 2, 00:06:55.719 "num_base_bdevs_discovered": 0, 00:06:55.719 "num_base_bdevs_operational": 2, 00:06:55.719 "base_bdevs_list": [ 00:06:55.719 { 00:06:55.719 "name": "BaseBdev1", 00:06:55.719 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:55.719 "is_configured": false, 00:06:55.719 "data_offset": 0, 00:06:55.719 "data_size": 0 00:06:55.719 }, 00:06:55.719 { 00:06:55.719 "name": "BaseBdev2", 00:06:55.719 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:55.719 "is_configured": false, 00:06:55.719 "data_offset": 0, 00:06:55.719 "data_size": 0 00:06:55.719 } 00:06:55.719 ] 00:06:55.719 }' 00:06:55.719 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:55.719 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.291 [2024-12-07 05:34:29.419218] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:06:56.291 [2024-12-07 05:34:29.419266] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.291 [2024-12-07 05:34:29.427196] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:56.291 [2024-12-07 05:34:29.427243] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:56.291 [2024-12-07 05:34:29.427252] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:56.291 [2024-12-07 05:34:29.427271] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.291 [2024-12-07 05:34:29.443970] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:56.291 BaseBdev1 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.291 [ 00:06:56.291 { 00:06:56.291 "name": "BaseBdev1", 00:06:56.291 "aliases": [ 00:06:56.291 "4e465880-a98b-4041-bdd3-00154124b59c" 00:06:56.291 ], 00:06:56.291 "product_name": "Malloc disk", 00:06:56.291 "block_size": 512, 00:06:56.291 "num_blocks": 65536, 00:06:56.291 "uuid": "4e465880-a98b-4041-bdd3-00154124b59c", 00:06:56.291 "assigned_rate_limits": { 00:06:56.291 "rw_ios_per_sec": 0, 00:06:56.291 "rw_mbytes_per_sec": 0, 00:06:56.291 "r_mbytes_per_sec": 0, 00:06:56.291 "w_mbytes_per_sec": 0 00:06:56.291 }, 00:06:56.291 "claimed": true, 00:06:56.291 "claim_type": "exclusive_write", 00:06:56.291 "zoned": false, 00:06:56.291 "supported_io_types": { 00:06:56.291 "read": true, 00:06:56.291 "write": true, 00:06:56.291 "unmap": true, 00:06:56.291 "flush": true, 00:06:56.291 "reset": true, 00:06:56.291 "nvme_admin": false, 00:06:56.291 "nvme_io": false, 00:06:56.291 "nvme_io_md": false, 00:06:56.291 "write_zeroes": true, 00:06:56.291 "zcopy": true, 00:06:56.291 "get_zone_info": false, 00:06:56.291 "zone_management": false, 00:06:56.291 "zone_append": false, 00:06:56.291 "compare": false, 00:06:56.291 "compare_and_write": false, 00:06:56.291 "abort": true, 00:06:56.291 "seek_hole": false, 00:06:56.291 "seek_data": false, 00:06:56.291 "copy": true, 00:06:56.291 "nvme_iov_md": false 00:06:56.291 }, 00:06:56.291 "memory_domains": [ 00:06:56.291 { 00:06:56.291 "dma_device_id": "system", 00:06:56.291 "dma_device_type": 1 00:06:56.291 }, 00:06:56.291 { 00:06:56.291 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:56.291 "dma_device_type": 2 00:06:56.291 } 00:06:56.291 ], 00:06:56.291 "driver_specific": {} 00:06:56.291 } 00:06:56.291 ] 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:56.291 "name": "Existed_Raid", 00:06:56.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:56.291 "strip_size_kb": 64, 00:06:56.291 "state": "configuring", 00:06:56.291 "raid_level": "raid0", 00:06:56.291 "superblock": false, 00:06:56.291 "num_base_bdevs": 2, 00:06:56.291 "num_base_bdevs_discovered": 1, 00:06:56.291 "num_base_bdevs_operational": 2, 00:06:56.291 "base_bdevs_list": [ 00:06:56.291 { 00:06:56.291 "name": "BaseBdev1", 00:06:56.291 "uuid": "4e465880-a98b-4041-bdd3-00154124b59c", 00:06:56.291 "is_configured": true, 00:06:56.291 "data_offset": 0, 00:06:56.291 "data_size": 65536 00:06:56.291 }, 00:06:56.291 { 00:06:56.291 "name": "BaseBdev2", 00:06:56.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:56.291 "is_configured": false, 00:06:56.291 "data_offset": 0, 00:06:56.291 "data_size": 0 00:06:56.291 } 00:06:56.291 ] 00:06:56.291 }' 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:56.291 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.552 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:56.552 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.552 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.552 [2024-12-07 05:34:29.827334] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:06:56.552 [2024-12-07 05:34:29.827377] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:06:56.552 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:56.552 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:56.552 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.552 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.552 [2024-12-07 05:34:29.835352] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:56.552 [2024-12-07 05:34:29.837141] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:56.552 [2024-12-07 05:34:29.837177] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:56.552 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:56.552 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:06:56.552 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:56.553 "name": "Existed_Raid", 00:06:56.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:56.553 "strip_size_kb": 64, 00:06:56.553 "state": "configuring", 00:06:56.553 "raid_level": "raid0", 00:06:56.553 "superblock": false, 00:06:56.553 "num_base_bdevs": 2, 00:06:56.553 "num_base_bdevs_discovered": 1, 00:06:56.553 "num_base_bdevs_operational": 2, 00:06:56.553 "base_bdevs_list": [ 00:06:56.553 { 00:06:56.553 "name": "BaseBdev1", 00:06:56.553 "uuid": "4e465880-a98b-4041-bdd3-00154124b59c", 00:06:56.553 "is_configured": true, 00:06:56.553 "data_offset": 0, 00:06:56.553 "data_size": 65536 00:06:56.553 }, 00:06:56.553 { 00:06:56.553 "name": "BaseBdev2", 00:06:56.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:56.553 "is_configured": false, 00:06:56.553 "data_offset": 0, 00:06:56.553 "data_size": 0 00:06:56.553 } 00:06:56.553 ] 00:06:56.553 }' 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:56.553 05:34:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.123 [2024-12-07 05:34:30.225630] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:06:57.123 [2024-12-07 05:34:30.225755] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:06:57.123 [2024-12-07 05:34:30.225793] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:57.123 [2024-12-07 05:34:30.226130] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:57.123 [2024-12-07 05:34:30.226345] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:06:57.123 [2024-12-07 05:34:30.226406] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:06:57.123 [2024-12-07 05:34:30.226646] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:57.123 BaseBdev2 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.123 [ 00:06:57.123 { 00:06:57.123 "name": "BaseBdev2", 00:06:57.123 "aliases": [ 00:06:57.123 "f79361a4-d89b-4de8-82e8-7c08e7826148" 00:06:57.123 ], 00:06:57.123 "product_name": "Malloc disk", 00:06:57.123 "block_size": 512, 00:06:57.123 "num_blocks": 65536, 00:06:57.123 "uuid": "f79361a4-d89b-4de8-82e8-7c08e7826148", 00:06:57.123 "assigned_rate_limits": { 00:06:57.123 "rw_ios_per_sec": 0, 00:06:57.123 "rw_mbytes_per_sec": 0, 00:06:57.123 "r_mbytes_per_sec": 0, 00:06:57.123 "w_mbytes_per_sec": 0 00:06:57.123 }, 00:06:57.123 "claimed": true, 00:06:57.123 "claim_type": "exclusive_write", 00:06:57.123 "zoned": false, 00:06:57.123 "supported_io_types": { 00:06:57.123 "read": true, 00:06:57.123 "write": true, 00:06:57.123 "unmap": true, 00:06:57.123 "flush": true, 00:06:57.123 "reset": true, 00:06:57.123 "nvme_admin": false, 00:06:57.123 "nvme_io": false, 00:06:57.123 "nvme_io_md": false, 00:06:57.123 "write_zeroes": true, 00:06:57.123 "zcopy": true, 00:06:57.123 "get_zone_info": false, 00:06:57.123 "zone_management": false, 00:06:57.123 "zone_append": false, 00:06:57.123 "compare": false, 00:06:57.123 "compare_and_write": false, 00:06:57.123 "abort": true, 00:06:57.123 "seek_hole": false, 00:06:57.123 "seek_data": false, 00:06:57.123 "copy": true, 00:06:57.123 "nvme_iov_md": false 00:06:57.123 }, 00:06:57.123 "memory_domains": [ 00:06:57.123 { 00:06:57.123 "dma_device_id": "system", 00:06:57.123 "dma_device_type": 1 00:06:57.123 }, 00:06:57.123 { 00:06:57.123 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:57.123 "dma_device_type": 2 00:06:57.123 } 00:06:57.123 ], 00:06:57.123 "driver_specific": {} 00:06:57.123 } 00:06:57.123 ] 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.123 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.124 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.124 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:57.124 "name": "Existed_Raid", 00:06:57.124 "uuid": "8a072496-50b3-45fb-a104-0a81a2007060", 00:06:57.124 "strip_size_kb": 64, 00:06:57.124 "state": "online", 00:06:57.124 "raid_level": "raid0", 00:06:57.124 "superblock": false, 00:06:57.124 "num_base_bdevs": 2, 00:06:57.124 "num_base_bdevs_discovered": 2, 00:06:57.124 "num_base_bdevs_operational": 2, 00:06:57.124 "base_bdevs_list": [ 00:06:57.124 { 00:06:57.124 "name": "BaseBdev1", 00:06:57.124 "uuid": "4e465880-a98b-4041-bdd3-00154124b59c", 00:06:57.124 "is_configured": true, 00:06:57.124 "data_offset": 0, 00:06:57.124 "data_size": 65536 00:06:57.124 }, 00:06:57.124 { 00:06:57.124 "name": "BaseBdev2", 00:06:57.124 "uuid": "f79361a4-d89b-4de8-82e8-7c08e7826148", 00:06:57.124 "is_configured": true, 00:06:57.124 "data_offset": 0, 00:06:57.124 "data_size": 65536 00:06:57.124 } 00:06:57.124 ] 00:06:57.124 }' 00:06:57.124 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:57.124 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.384 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:06:57.384 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:06:57.384 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:06:57.384 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:06:57.384 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:06:57.384 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:06:57.384 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:06:57.384 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:06:57.384 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.385 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.645 [2024-12-07 05:34:30.756996] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:06:57.645 "name": "Existed_Raid", 00:06:57.645 "aliases": [ 00:06:57.645 "8a072496-50b3-45fb-a104-0a81a2007060" 00:06:57.645 ], 00:06:57.645 "product_name": "Raid Volume", 00:06:57.645 "block_size": 512, 00:06:57.645 "num_blocks": 131072, 00:06:57.645 "uuid": "8a072496-50b3-45fb-a104-0a81a2007060", 00:06:57.645 "assigned_rate_limits": { 00:06:57.645 "rw_ios_per_sec": 0, 00:06:57.645 "rw_mbytes_per_sec": 0, 00:06:57.645 "r_mbytes_per_sec": 0, 00:06:57.645 "w_mbytes_per_sec": 0 00:06:57.645 }, 00:06:57.645 "claimed": false, 00:06:57.645 "zoned": false, 00:06:57.645 "supported_io_types": { 00:06:57.645 "read": true, 00:06:57.645 "write": true, 00:06:57.645 "unmap": true, 00:06:57.645 "flush": true, 00:06:57.645 "reset": true, 00:06:57.645 "nvme_admin": false, 00:06:57.645 "nvme_io": false, 00:06:57.645 "nvme_io_md": false, 00:06:57.645 "write_zeroes": true, 00:06:57.645 "zcopy": false, 00:06:57.645 "get_zone_info": false, 00:06:57.645 "zone_management": false, 00:06:57.645 "zone_append": false, 00:06:57.645 "compare": false, 00:06:57.645 "compare_and_write": false, 00:06:57.645 "abort": false, 00:06:57.645 "seek_hole": false, 00:06:57.645 "seek_data": false, 00:06:57.645 "copy": false, 00:06:57.645 "nvme_iov_md": false 00:06:57.645 }, 00:06:57.645 "memory_domains": [ 00:06:57.645 { 00:06:57.645 "dma_device_id": "system", 00:06:57.645 "dma_device_type": 1 00:06:57.645 }, 00:06:57.645 { 00:06:57.645 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:57.645 "dma_device_type": 2 00:06:57.645 }, 00:06:57.645 { 00:06:57.645 "dma_device_id": "system", 00:06:57.645 "dma_device_type": 1 00:06:57.645 }, 00:06:57.645 { 00:06:57.645 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:57.645 "dma_device_type": 2 00:06:57.645 } 00:06:57.645 ], 00:06:57.645 "driver_specific": { 00:06:57.645 "raid": { 00:06:57.645 "uuid": "8a072496-50b3-45fb-a104-0a81a2007060", 00:06:57.645 "strip_size_kb": 64, 00:06:57.645 "state": "online", 00:06:57.645 "raid_level": "raid0", 00:06:57.645 "superblock": false, 00:06:57.645 "num_base_bdevs": 2, 00:06:57.645 "num_base_bdevs_discovered": 2, 00:06:57.645 "num_base_bdevs_operational": 2, 00:06:57.645 "base_bdevs_list": [ 00:06:57.645 { 00:06:57.645 "name": "BaseBdev1", 00:06:57.645 "uuid": "4e465880-a98b-4041-bdd3-00154124b59c", 00:06:57.645 "is_configured": true, 00:06:57.645 "data_offset": 0, 00:06:57.645 "data_size": 65536 00:06:57.645 }, 00:06:57.645 { 00:06:57.645 "name": "BaseBdev2", 00:06:57.645 "uuid": "f79361a4-d89b-4de8-82e8-7c08e7826148", 00:06:57.645 "is_configured": true, 00:06:57.645 "data_offset": 0, 00:06:57.645 "data_size": 65536 00:06:57.645 } 00:06:57.645 ] 00:06:57.645 } 00:06:57.645 } 00:06:57.645 }' 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:06:57.645 BaseBdev2' 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.645 05:34:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.645 [2024-12-07 05:34:31.000365] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:06:57.645 [2024-12-07 05:34:31.000392] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:57.645 [2024-12-07 05:34:31.000442] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:57.905 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:57.905 "name": "Existed_Raid", 00:06:57.906 "uuid": "8a072496-50b3-45fb-a104-0a81a2007060", 00:06:57.906 "strip_size_kb": 64, 00:06:57.906 "state": "offline", 00:06:57.906 "raid_level": "raid0", 00:06:57.906 "superblock": false, 00:06:57.906 "num_base_bdevs": 2, 00:06:57.906 "num_base_bdevs_discovered": 1, 00:06:57.906 "num_base_bdevs_operational": 1, 00:06:57.906 "base_bdevs_list": [ 00:06:57.906 { 00:06:57.906 "name": null, 00:06:57.906 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:57.906 "is_configured": false, 00:06:57.906 "data_offset": 0, 00:06:57.906 "data_size": 65536 00:06:57.906 }, 00:06:57.906 { 00:06:57.906 "name": "BaseBdev2", 00:06:57.906 "uuid": "f79361a4-d89b-4de8-82e8-7c08e7826148", 00:06:57.906 "is_configured": true, 00:06:57.906 "data_offset": 0, 00:06:57.906 "data_size": 65536 00:06:57.906 } 00:06:57.906 ] 00:06:57.906 }' 00:06:57.906 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:57.906 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.165 [2024-12-07 05:34:31.406875] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:06:58.165 [2024-12-07 05:34:31.406973] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:58.165 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 71813 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 71813 ']' 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 71813 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71813 00:06:58.166 killing process with pid 71813 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71813' 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 71813 00:06:58.166 [2024-12-07 05:34:31.514970] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:58.166 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 71813 00:06:58.166 [2024-12-07 05:34:31.515960] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:58.425 05:34:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:06:58.425 00:06:58.425 real 0m3.697s 00:06:58.425 user 0m5.861s 00:06:58.425 sys 0m0.695s 00:06:58.425 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:58.425 05:34:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.425 ************************************ 00:06:58.425 END TEST raid_state_function_test 00:06:58.425 ************************************ 00:06:58.425 05:34:31 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:06:58.425 05:34:31 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:58.425 05:34:31 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:58.425 05:34:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:58.685 ************************************ 00:06:58.685 START TEST raid_state_function_test_sb 00:06:58.685 ************************************ 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 true 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72050 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72050' 00:06:58.685 Process raid pid: 72050 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72050 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 72050 ']' 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:58.685 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:58.685 05:34:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:58.685 [2024-12-07 05:34:31.884937] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:06:58.685 [2024-12-07 05:34:31.885149] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:58.685 [2024-12-07 05:34:32.035081] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:58.944 [2024-12-07 05:34:32.059340] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:58.944 [2024-12-07 05:34:32.100807] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:58.944 [2024-12-07 05:34:32.100921] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:59.514 [2024-12-07 05:34:32.707000] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:59.514 [2024-12-07 05:34:32.707060] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:59.514 [2024-12-07 05:34:32.707070] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:59.514 [2024-12-07 05:34:32.707079] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:59.514 "name": "Existed_Raid", 00:06:59.514 "uuid": "c266f932-4d75-4e81-b7bc-e45db9bf768f", 00:06:59.514 "strip_size_kb": 64, 00:06:59.514 "state": "configuring", 00:06:59.514 "raid_level": "raid0", 00:06:59.514 "superblock": true, 00:06:59.514 "num_base_bdevs": 2, 00:06:59.514 "num_base_bdevs_discovered": 0, 00:06:59.514 "num_base_bdevs_operational": 2, 00:06:59.514 "base_bdevs_list": [ 00:06:59.514 { 00:06:59.514 "name": "BaseBdev1", 00:06:59.514 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:59.514 "is_configured": false, 00:06:59.514 "data_offset": 0, 00:06:59.514 "data_size": 0 00:06:59.514 }, 00:06:59.514 { 00:06:59.514 "name": "BaseBdev2", 00:06:59.514 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:59.514 "is_configured": false, 00:06:59.514 "data_offset": 0, 00:06:59.514 "data_size": 0 00:06:59.514 } 00:06:59.514 ] 00:06:59.514 }' 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:59.514 05:34:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:06:59.774 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:59.774 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:59.774 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.034 [2024-12-07 05:34:33.142165] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:00.034 [2024-12-07 05:34:33.142256] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.034 [2024-12-07 05:34:33.154164] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:00.034 [2024-12-07 05:34:33.154248] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:00.034 [2024-12-07 05:34:33.154275] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:00.034 [2024-12-07 05:34:33.154311] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.034 [2024-12-07 05:34:33.175115] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:00.034 BaseBdev1 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.034 [ 00:07:00.034 { 00:07:00.034 "name": "BaseBdev1", 00:07:00.034 "aliases": [ 00:07:00.034 "d89deb60-7461-4572-9e25-35bf64e15c59" 00:07:00.034 ], 00:07:00.034 "product_name": "Malloc disk", 00:07:00.034 "block_size": 512, 00:07:00.034 "num_blocks": 65536, 00:07:00.034 "uuid": "d89deb60-7461-4572-9e25-35bf64e15c59", 00:07:00.034 "assigned_rate_limits": { 00:07:00.034 "rw_ios_per_sec": 0, 00:07:00.034 "rw_mbytes_per_sec": 0, 00:07:00.034 "r_mbytes_per_sec": 0, 00:07:00.034 "w_mbytes_per_sec": 0 00:07:00.034 }, 00:07:00.034 "claimed": true, 00:07:00.034 "claim_type": "exclusive_write", 00:07:00.034 "zoned": false, 00:07:00.034 "supported_io_types": { 00:07:00.034 "read": true, 00:07:00.034 "write": true, 00:07:00.034 "unmap": true, 00:07:00.034 "flush": true, 00:07:00.034 "reset": true, 00:07:00.034 "nvme_admin": false, 00:07:00.034 "nvme_io": false, 00:07:00.034 "nvme_io_md": false, 00:07:00.034 "write_zeroes": true, 00:07:00.034 "zcopy": true, 00:07:00.034 "get_zone_info": false, 00:07:00.034 "zone_management": false, 00:07:00.034 "zone_append": false, 00:07:00.034 "compare": false, 00:07:00.034 "compare_and_write": false, 00:07:00.034 "abort": true, 00:07:00.034 "seek_hole": false, 00:07:00.034 "seek_data": false, 00:07:00.034 "copy": true, 00:07:00.034 "nvme_iov_md": false 00:07:00.034 }, 00:07:00.034 "memory_domains": [ 00:07:00.034 { 00:07:00.034 "dma_device_id": "system", 00:07:00.034 "dma_device_type": 1 00:07:00.034 }, 00:07:00.034 { 00:07:00.034 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:00.034 "dma_device_type": 2 00:07:00.034 } 00:07:00.034 ], 00:07:00.034 "driver_specific": {} 00:07:00.034 } 00:07:00.034 ] 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:00.034 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:00.035 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:00.035 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:00.035 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:00.035 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:00.035 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:00.035 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.035 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.035 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.035 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:00.035 "name": "Existed_Raid", 00:07:00.035 "uuid": "cf0d4f6d-e4b4-4d20-8c00-19066182bd8e", 00:07:00.035 "strip_size_kb": 64, 00:07:00.035 "state": "configuring", 00:07:00.035 "raid_level": "raid0", 00:07:00.035 "superblock": true, 00:07:00.035 "num_base_bdevs": 2, 00:07:00.035 "num_base_bdevs_discovered": 1, 00:07:00.035 "num_base_bdevs_operational": 2, 00:07:00.035 "base_bdevs_list": [ 00:07:00.035 { 00:07:00.035 "name": "BaseBdev1", 00:07:00.035 "uuid": "d89deb60-7461-4572-9e25-35bf64e15c59", 00:07:00.035 "is_configured": true, 00:07:00.035 "data_offset": 2048, 00:07:00.035 "data_size": 63488 00:07:00.035 }, 00:07:00.035 { 00:07:00.035 "name": "BaseBdev2", 00:07:00.035 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.035 "is_configured": false, 00:07:00.035 "data_offset": 0, 00:07:00.035 "data_size": 0 00:07:00.035 } 00:07:00.035 ] 00:07:00.035 }' 00:07:00.035 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:00.035 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.603 [2024-12-07 05:34:33.666344] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:00.603 [2024-12-07 05:34:33.666432] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.603 [2024-12-07 05:34:33.674356] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:00.603 [2024-12-07 05:34:33.676205] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:00.603 [2024-12-07 05:34:33.676248] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:00.603 "name": "Existed_Raid", 00:07:00.603 "uuid": "9e69b02d-60fa-4206-85b6-0d4ca8acbbe3", 00:07:00.603 "strip_size_kb": 64, 00:07:00.603 "state": "configuring", 00:07:00.603 "raid_level": "raid0", 00:07:00.603 "superblock": true, 00:07:00.603 "num_base_bdevs": 2, 00:07:00.603 "num_base_bdevs_discovered": 1, 00:07:00.603 "num_base_bdevs_operational": 2, 00:07:00.603 "base_bdevs_list": [ 00:07:00.603 { 00:07:00.603 "name": "BaseBdev1", 00:07:00.603 "uuid": "d89deb60-7461-4572-9e25-35bf64e15c59", 00:07:00.603 "is_configured": true, 00:07:00.603 "data_offset": 2048, 00:07:00.603 "data_size": 63488 00:07:00.603 }, 00:07:00.603 { 00:07:00.603 "name": "BaseBdev2", 00:07:00.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.603 "is_configured": false, 00:07:00.603 "data_offset": 0, 00:07:00.603 "data_size": 0 00:07:00.603 } 00:07:00.603 ] 00:07:00.603 }' 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:00.603 05:34:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.863 [2024-12-07 05:34:34.056459] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:00.863 [2024-12-07 05:34:34.056661] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:00.863 [2024-12-07 05:34:34.056677] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:00.863 BaseBdev2 00:07:00.863 [2024-12-07 05:34:34.056970] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:00.863 [2024-12-07 05:34:34.057116] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:00.863 [2024-12-07 05:34:34.057135] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:00.863 [2024-12-07 05:34:34.057250] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.863 [ 00:07:00.863 { 00:07:00.863 "name": "BaseBdev2", 00:07:00.863 "aliases": [ 00:07:00.863 "f6dc8470-7653-4719-bbda-78bd9004bb7f" 00:07:00.863 ], 00:07:00.863 "product_name": "Malloc disk", 00:07:00.863 "block_size": 512, 00:07:00.863 "num_blocks": 65536, 00:07:00.863 "uuid": "f6dc8470-7653-4719-bbda-78bd9004bb7f", 00:07:00.863 "assigned_rate_limits": { 00:07:00.863 "rw_ios_per_sec": 0, 00:07:00.863 "rw_mbytes_per_sec": 0, 00:07:00.863 "r_mbytes_per_sec": 0, 00:07:00.863 "w_mbytes_per_sec": 0 00:07:00.863 }, 00:07:00.863 "claimed": true, 00:07:00.863 "claim_type": "exclusive_write", 00:07:00.863 "zoned": false, 00:07:00.863 "supported_io_types": { 00:07:00.863 "read": true, 00:07:00.863 "write": true, 00:07:00.863 "unmap": true, 00:07:00.863 "flush": true, 00:07:00.863 "reset": true, 00:07:00.863 "nvme_admin": false, 00:07:00.863 "nvme_io": false, 00:07:00.863 "nvme_io_md": false, 00:07:00.863 "write_zeroes": true, 00:07:00.863 "zcopy": true, 00:07:00.863 "get_zone_info": false, 00:07:00.863 "zone_management": false, 00:07:00.863 "zone_append": false, 00:07:00.863 "compare": false, 00:07:00.863 "compare_and_write": false, 00:07:00.863 "abort": true, 00:07:00.863 "seek_hole": false, 00:07:00.863 "seek_data": false, 00:07:00.863 "copy": true, 00:07:00.863 "nvme_iov_md": false 00:07:00.863 }, 00:07:00.863 "memory_domains": [ 00:07:00.863 { 00:07:00.863 "dma_device_id": "system", 00:07:00.863 "dma_device_type": 1 00:07:00.863 }, 00:07:00.863 { 00:07:00.863 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:00.863 "dma_device_type": 2 00:07:00.863 } 00:07:00.863 ], 00:07:00.863 "driver_specific": {} 00:07:00.863 } 00:07:00.863 ] 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:00.863 "name": "Existed_Raid", 00:07:00.863 "uuid": "9e69b02d-60fa-4206-85b6-0d4ca8acbbe3", 00:07:00.863 "strip_size_kb": 64, 00:07:00.863 "state": "online", 00:07:00.863 "raid_level": "raid0", 00:07:00.863 "superblock": true, 00:07:00.863 "num_base_bdevs": 2, 00:07:00.863 "num_base_bdevs_discovered": 2, 00:07:00.863 "num_base_bdevs_operational": 2, 00:07:00.863 "base_bdevs_list": [ 00:07:00.863 { 00:07:00.863 "name": "BaseBdev1", 00:07:00.863 "uuid": "d89deb60-7461-4572-9e25-35bf64e15c59", 00:07:00.863 "is_configured": true, 00:07:00.863 "data_offset": 2048, 00:07:00.863 "data_size": 63488 00:07:00.863 }, 00:07:00.863 { 00:07:00.863 "name": "BaseBdev2", 00:07:00.863 "uuid": "f6dc8470-7653-4719-bbda-78bd9004bb7f", 00:07:00.863 "is_configured": true, 00:07:00.863 "data_offset": 2048, 00:07:00.863 "data_size": 63488 00:07:00.863 } 00:07:00.863 ] 00:07:00.863 }' 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:00.863 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:01.432 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:01.432 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:01.433 [2024-12-07 05:34:34.531937] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:01.433 "name": "Existed_Raid", 00:07:01.433 "aliases": [ 00:07:01.433 "9e69b02d-60fa-4206-85b6-0d4ca8acbbe3" 00:07:01.433 ], 00:07:01.433 "product_name": "Raid Volume", 00:07:01.433 "block_size": 512, 00:07:01.433 "num_blocks": 126976, 00:07:01.433 "uuid": "9e69b02d-60fa-4206-85b6-0d4ca8acbbe3", 00:07:01.433 "assigned_rate_limits": { 00:07:01.433 "rw_ios_per_sec": 0, 00:07:01.433 "rw_mbytes_per_sec": 0, 00:07:01.433 "r_mbytes_per_sec": 0, 00:07:01.433 "w_mbytes_per_sec": 0 00:07:01.433 }, 00:07:01.433 "claimed": false, 00:07:01.433 "zoned": false, 00:07:01.433 "supported_io_types": { 00:07:01.433 "read": true, 00:07:01.433 "write": true, 00:07:01.433 "unmap": true, 00:07:01.433 "flush": true, 00:07:01.433 "reset": true, 00:07:01.433 "nvme_admin": false, 00:07:01.433 "nvme_io": false, 00:07:01.433 "nvme_io_md": false, 00:07:01.433 "write_zeroes": true, 00:07:01.433 "zcopy": false, 00:07:01.433 "get_zone_info": false, 00:07:01.433 "zone_management": false, 00:07:01.433 "zone_append": false, 00:07:01.433 "compare": false, 00:07:01.433 "compare_and_write": false, 00:07:01.433 "abort": false, 00:07:01.433 "seek_hole": false, 00:07:01.433 "seek_data": false, 00:07:01.433 "copy": false, 00:07:01.433 "nvme_iov_md": false 00:07:01.433 }, 00:07:01.433 "memory_domains": [ 00:07:01.433 { 00:07:01.433 "dma_device_id": "system", 00:07:01.433 "dma_device_type": 1 00:07:01.433 }, 00:07:01.433 { 00:07:01.433 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:01.433 "dma_device_type": 2 00:07:01.433 }, 00:07:01.433 { 00:07:01.433 "dma_device_id": "system", 00:07:01.433 "dma_device_type": 1 00:07:01.433 }, 00:07:01.433 { 00:07:01.433 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:01.433 "dma_device_type": 2 00:07:01.433 } 00:07:01.433 ], 00:07:01.433 "driver_specific": { 00:07:01.433 "raid": { 00:07:01.433 "uuid": "9e69b02d-60fa-4206-85b6-0d4ca8acbbe3", 00:07:01.433 "strip_size_kb": 64, 00:07:01.433 "state": "online", 00:07:01.433 "raid_level": "raid0", 00:07:01.433 "superblock": true, 00:07:01.433 "num_base_bdevs": 2, 00:07:01.433 "num_base_bdevs_discovered": 2, 00:07:01.433 "num_base_bdevs_operational": 2, 00:07:01.433 "base_bdevs_list": [ 00:07:01.433 { 00:07:01.433 "name": "BaseBdev1", 00:07:01.433 "uuid": "d89deb60-7461-4572-9e25-35bf64e15c59", 00:07:01.433 "is_configured": true, 00:07:01.433 "data_offset": 2048, 00:07:01.433 "data_size": 63488 00:07:01.433 }, 00:07:01.433 { 00:07:01.433 "name": "BaseBdev2", 00:07:01.433 "uuid": "f6dc8470-7653-4719-bbda-78bd9004bb7f", 00:07:01.433 "is_configured": true, 00:07:01.433 "data_offset": 2048, 00:07:01.433 "data_size": 63488 00:07:01.433 } 00:07:01.433 ] 00:07:01.433 } 00:07:01.433 } 00:07:01.433 }' 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:01.433 BaseBdev2' 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:01.433 [2024-12-07 05:34:34.779291] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:01.433 [2024-12-07 05:34:34.779364] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:01.433 [2024-12-07 05:34:34.779432] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:01.433 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:01.434 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:01.693 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:01.693 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:01.693 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.693 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:01.693 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.693 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:01.693 "name": "Existed_Raid", 00:07:01.693 "uuid": "9e69b02d-60fa-4206-85b6-0d4ca8acbbe3", 00:07:01.693 "strip_size_kb": 64, 00:07:01.693 "state": "offline", 00:07:01.693 "raid_level": "raid0", 00:07:01.693 "superblock": true, 00:07:01.693 "num_base_bdevs": 2, 00:07:01.693 "num_base_bdevs_discovered": 1, 00:07:01.693 "num_base_bdevs_operational": 1, 00:07:01.693 "base_bdevs_list": [ 00:07:01.693 { 00:07:01.693 "name": null, 00:07:01.693 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:01.693 "is_configured": false, 00:07:01.693 "data_offset": 0, 00:07:01.693 "data_size": 63488 00:07:01.693 }, 00:07:01.693 { 00:07:01.693 "name": "BaseBdev2", 00:07:01.693 "uuid": "f6dc8470-7653-4719-bbda-78bd9004bb7f", 00:07:01.693 "is_configured": true, 00:07:01.693 "data_offset": 2048, 00:07:01.693 "data_size": 63488 00:07:01.693 } 00:07:01.693 ] 00:07:01.693 }' 00:07:01.693 05:34:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:01.693 05:34:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:01.955 [2024-12-07 05:34:35.257614] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:01.955 [2024-12-07 05:34:35.257715] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:01.955 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72050 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 72050 ']' 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 72050 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72050 00:07:02.215 killing process with pid 72050 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72050' 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 72050 00:07:02.215 [2024-12-07 05:34:35.364708] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 72050 00:07:02.215 [2024-12-07 05:34:35.365681] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:02.215 00:07:02.215 real 0m3.782s 00:07:02.215 user 0m6.004s 00:07:02.215 sys 0m0.719s 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:02.215 05:34:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:02.215 ************************************ 00:07:02.215 END TEST raid_state_function_test_sb 00:07:02.215 ************************************ 00:07:02.475 05:34:35 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:07:02.475 05:34:35 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:02.475 05:34:35 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:02.475 05:34:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:02.475 ************************************ 00:07:02.475 START TEST raid_superblock_test 00:07:02.475 ************************************ 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 2 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=72285 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 72285 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 72285 ']' 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:02.475 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:02.475 05:34:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:02.475 [2024-12-07 05:34:35.727443] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:02.475 [2024-12-07 05:34:35.727557] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72285 ] 00:07:02.734 [2024-12-07 05:34:35.880700] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:02.734 [2024-12-07 05:34:35.905330] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:02.734 [2024-12-07 05:34:35.946787] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:02.734 [2024-12-07 05:34:35.946823] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.304 malloc1 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.304 [2024-12-07 05:34:36.573561] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:03.304 [2024-12-07 05:34:36.573701] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:03.304 [2024-12-07 05:34:36.573748] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:03.304 [2024-12-07 05:34:36.573792] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:03.304 [2024-12-07 05:34:36.575912] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:03.304 [2024-12-07 05:34:36.575999] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:03.304 pt1 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.304 malloc2 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.304 [2024-12-07 05:34:36.605869] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:03.304 [2024-12-07 05:34:36.605924] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:03.304 [2024-12-07 05:34:36.605941] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:03.304 [2024-12-07 05:34:36.605951] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:03.304 [2024-12-07 05:34:36.607993] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:03.304 [2024-12-07 05:34:36.608031] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:03.304 pt2 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.304 [2024-12-07 05:34:36.617878] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:03.304 [2024-12-07 05:34:36.619742] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:03.304 [2024-12-07 05:34:36.619881] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:03.304 [2024-12-07 05:34:36.619900] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:03.304 [2024-12-07 05:34:36.620169] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:03.304 [2024-12-07 05:34:36.620318] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:03.304 [2024-12-07 05:34:36.620328] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:03.304 [2024-12-07 05:34:36.620444] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:03.304 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.305 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:03.564 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:03.564 "name": "raid_bdev1", 00:07:03.564 "uuid": "e46bb084-fd0e-49e5-b5c0-3fb0382282b0", 00:07:03.564 "strip_size_kb": 64, 00:07:03.564 "state": "online", 00:07:03.564 "raid_level": "raid0", 00:07:03.564 "superblock": true, 00:07:03.564 "num_base_bdevs": 2, 00:07:03.564 "num_base_bdevs_discovered": 2, 00:07:03.564 "num_base_bdevs_operational": 2, 00:07:03.564 "base_bdevs_list": [ 00:07:03.564 { 00:07:03.564 "name": "pt1", 00:07:03.564 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:03.564 "is_configured": true, 00:07:03.564 "data_offset": 2048, 00:07:03.564 "data_size": 63488 00:07:03.564 }, 00:07:03.564 { 00:07:03.564 "name": "pt2", 00:07:03.564 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:03.564 "is_configured": true, 00:07:03.564 "data_offset": 2048, 00:07:03.564 "data_size": 63488 00:07:03.564 } 00:07:03.564 ] 00:07:03.564 }' 00:07:03.564 05:34:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:03.564 05:34:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.823 [2024-12-07 05:34:37.077368] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:03.823 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:03.823 "name": "raid_bdev1", 00:07:03.823 "aliases": [ 00:07:03.823 "e46bb084-fd0e-49e5-b5c0-3fb0382282b0" 00:07:03.823 ], 00:07:03.823 "product_name": "Raid Volume", 00:07:03.823 "block_size": 512, 00:07:03.823 "num_blocks": 126976, 00:07:03.823 "uuid": "e46bb084-fd0e-49e5-b5c0-3fb0382282b0", 00:07:03.823 "assigned_rate_limits": { 00:07:03.824 "rw_ios_per_sec": 0, 00:07:03.824 "rw_mbytes_per_sec": 0, 00:07:03.824 "r_mbytes_per_sec": 0, 00:07:03.824 "w_mbytes_per_sec": 0 00:07:03.824 }, 00:07:03.824 "claimed": false, 00:07:03.824 "zoned": false, 00:07:03.824 "supported_io_types": { 00:07:03.824 "read": true, 00:07:03.824 "write": true, 00:07:03.824 "unmap": true, 00:07:03.824 "flush": true, 00:07:03.824 "reset": true, 00:07:03.824 "nvme_admin": false, 00:07:03.824 "nvme_io": false, 00:07:03.824 "nvme_io_md": false, 00:07:03.824 "write_zeroes": true, 00:07:03.824 "zcopy": false, 00:07:03.824 "get_zone_info": false, 00:07:03.824 "zone_management": false, 00:07:03.824 "zone_append": false, 00:07:03.824 "compare": false, 00:07:03.824 "compare_and_write": false, 00:07:03.824 "abort": false, 00:07:03.824 "seek_hole": false, 00:07:03.824 "seek_data": false, 00:07:03.824 "copy": false, 00:07:03.824 "nvme_iov_md": false 00:07:03.824 }, 00:07:03.824 "memory_domains": [ 00:07:03.824 { 00:07:03.824 "dma_device_id": "system", 00:07:03.824 "dma_device_type": 1 00:07:03.824 }, 00:07:03.824 { 00:07:03.824 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:03.824 "dma_device_type": 2 00:07:03.824 }, 00:07:03.824 { 00:07:03.824 "dma_device_id": "system", 00:07:03.824 "dma_device_type": 1 00:07:03.824 }, 00:07:03.824 { 00:07:03.824 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:03.824 "dma_device_type": 2 00:07:03.824 } 00:07:03.824 ], 00:07:03.824 "driver_specific": { 00:07:03.824 "raid": { 00:07:03.824 "uuid": "e46bb084-fd0e-49e5-b5c0-3fb0382282b0", 00:07:03.824 "strip_size_kb": 64, 00:07:03.824 "state": "online", 00:07:03.824 "raid_level": "raid0", 00:07:03.824 "superblock": true, 00:07:03.824 "num_base_bdevs": 2, 00:07:03.824 "num_base_bdevs_discovered": 2, 00:07:03.824 "num_base_bdevs_operational": 2, 00:07:03.824 "base_bdevs_list": [ 00:07:03.824 { 00:07:03.824 "name": "pt1", 00:07:03.824 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:03.824 "is_configured": true, 00:07:03.824 "data_offset": 2048, 00:07:03.824 "data_size": 63488 00:07:03.824 }, 00:07:03.824 { 00:07:03.824 "name": "pt2", 00:07:03.824 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:03.824 "is_configured": true, 00:07:03.824 "data_offset": 2048, 00:07:03.824 "data_size": 63488 00:07:03.824 } 00:07:03.824 ] 00:07:03.824 } 00:07:03.824 } 00:07:03.824 }' 00:07:03.824 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:03.824 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:03.824 pt2' 00:07:03.824 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:03.824 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:03.824 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:03.824 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:03.824 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:03.824 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.824 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:04.090 [2024-12-07 05:34:37.292900] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=e46bb084-fd0e-49e5-b5c0-3fb0382282b0 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z e46bb084-fd0e-49e5-b5c0-3fb0382282b0 ']' 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.090 [2024-12-07 05:34:37.340593] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:04.090 [2024-12-07 05:34:37.340619] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:04.090 [2024-12-07 05:34:37.340697] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:04.090 [2024-12-07 05:34:37.340744] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:04.090 [2024-12-07 05:34:37.340760] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:04.090 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.091 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.367 [2024-12-07 05:34:37.476384] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:04.367 [2024-12-07 05:34:37.478346] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:04.367 [2024-12-07 05:34:37.478503] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:04.367 [2024-12-07 05:34:37.478598] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:04.367 [2024-12-07 05:34:37.478668] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:04.367 [2024-12-07 05:34:37.478721] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:04.367 request: 00:07:04.367 { 00:07:04.367 "name": "raid_bdev1", 00:07:04.367 "raid_level": "raid0", 00:07:04.367 "base_bdevs": [ 00:07:04.367 "malloc1", 00:07:04.367 "malloc2" 00:07:04.367 ], 00:07:04.367 "strip_size_kb": 64, 00:07:04.367 "superblock": false, 00:07:04.367 "method": "bdev_raid_create", 00:07:04.367 "req_id": 1 00:07:04.367 } 00:07:04.367 Got JSON-RPC error response 00:07:04.367 response: 00:07:04.367 { 00:07:04.367 "code": -17, 00:07:04.367 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:04.367 } 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:07:04.367 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.368 [2024-12-07 05:34:37.544280] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:04.368 [2024-12-07 05:34:37.544410] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:04.368 [2024-12-07 05:34:37.544456] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:04.368 [2024-12-07 05:34:37.544489] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:04.368 [2024-12-07 05:34:37.546746] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:04.368 [2024-12-07 05:34:37.546822] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:04.368 [2024-12-07 05:34:37.546937] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:04.368 [2024-12-07 05:34:37.547002] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:04.368 pt1 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:04.368 "name": "raid_bdev1", 00:07:04.368 "uuid": "e46bb084-fd0e-49e5-b5c0-3fb0382282b0", 00:07:04.368 "strip_size_kb": 64, 00:07:04.368 "state": "configuring", 00:07:04.368 "raid_level": "raid0", 00:07:04.368 "superblock": true, 00:07:04.368 "num_base_bdevs": 2, 00:07:04.368 "num_base_bdevs_discovered": 1, 00:07:04.368 "num_base_bdevs_operational": 2, 00:07:04.368 "base_bdevs_list": [ 00:07:04.368 { 00:07:04.368 "name": "pt1", 00:07:04.368 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:04.368 "is_configured": true, 00:07:04.368 "data_offset": 2048, 00:07:04.368 "data_size": 63488 00:07:04.368 }, 00:07:04.368 { 00:07:04.368 "name": null, 00:07:04.368 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:04.368 "is_configured": false, 00:07:04.368 "data_offset": 2048, 00:07:04.368 "data_size": 63488 00:07:04.368 } 00:07:04.368 ] 00:07:04.368 }' 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:04.368 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.626 [2024-12-07 05:34:37.971522] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:04.626 [2024-12-07 05:34:37.971639] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:04.626 [2024-12-07 05:34:37.971680] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:04.626 [2024-12-07 05:34:37.971709] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:04.626 [2024-12-07 05:34:37.972121] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:04.626 [2024-12-07 05:34:37.972180] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:04.626 [2024-12-07 05:34:37.972283] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:04.626 [2024-12-07 05:34:37.972331] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:04.626 [2024-12-07 05:34:37.972455] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:04.626 [2024-12-07 05:34:37.972490] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:04.626 [2024-12-07 05:34:37.972780] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:04.626 [2024-12-07 05:34:37.972927] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:04.626 [2024-12-07 05:34:37.972971] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:04.626 [2024-12-07 05:34:37.973113] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:04.626 pt2 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.626 05:34:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.885 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.885 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:04.885 "name": "raid_bdev1", 00:07:04.885 "uuid": "e46bb084-fd0e-49e5-b5c0-3fb0382282b0", 00:07:04.885 "strip_size_kb": 64, 00:07:04.885 "state": "online", 00:07:04.885 "raid_level": "raid0", 00:07:04.885 "superblock": true, 00:07:04.885 "num_base_bdevs": 2, 00:07:04.885 "num_base_bdevs_discovered": 2, 00:07:04.885 "num_base_bdevs_operational": 2, 00:07:04.885 "base_bdevs_list": [ 00:07:04.885 { 00:07:04.885 "name": "pt1", 00:07:04.885 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:04.885 "is_configured": true, 00:07:04.885 "data_offset": 2048, 00:07:04.885 "data_size": 63488 00:07:04.885 }, 00:07:04.885 { 00:07:04.885 "name": "pt2", 00:07:04.885 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:04.885 "is_configured": true, 00:07:04.885 "data_offset": 2048, 00:07:04.885 "data_size": 63488 00:07:04.885 } 00:07:04.885 ] 00:07:04.885 }' 00:07:04.885 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:04.885 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.183 [2024-12-07 05:34:38.407046] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:05.183 "name": "raid_bdev1", 00:07:05.183 "aliases": [ 00:07:05.183 "e46bb084-fd0e-49e5-b5c0-3fb0382282b0" 00:07:05.183 ], 00:07:05.183 "product_name": "Raid Volume", 00:07:05.183 "block_size": 512, 00:07:05.183 "num_blocks": 126976, 00:07:05.183 "uuid": "e46bb084-fd0e-49e5-b5c0-3fb0382282b0", 00:07:05.183 "assigned_rate_limits": { 00:07:05.183 "rw_ios_per_sec": 0, 00:07:05.183 "rw_mbytes_per_sec": 0, 00:07:05.183 "r_mbytes_per_sec": 0, 00:07:05.183 "w_mbytes_per_sec": 0 00:07:05.183 }, 00:07:05.183 "claimed": false, 00:07:05.183 "zoned": false, 00:07:05.183 "supported_io_types": { 00:07:05.183 "read": true, 00:07:05.183 "write": true, 00:07:05.183 "unmap": true, 00:07:05.183 "flush": true, 00:07:05.183 "reset": true, 00:07:05.183 "nvme_admin": false, 00:07:05.183 "nvme_io": false, 00:07:05.183 "nvme_io_md": false, 00:07:05.183 "write_zeroes": true, 00:07:05.183 "zcopy": false, 00:07:05.183 "get_zone_info": false, 00:07:05.183 "zone_management": false, 00:07:05.183 "zone_append": false, 00:07:05.183 "compare": false, 00:07:05.183 "compare_and_write": false, 00:07:05.183 "abort": false, 00:07:05.183 "seek_hole": false, 00:07:05.183 "seek_data": false, 00:07:05.183 "copy": false, 00:07:05.183 "nvme_iov_md": false 00:07:05.183 }, 00:07:05.183 "memory_domains": [ 00:07:05.183 { 00:07:05.183 "dma_device_id": "system", 00:07:05.183 "dma_device_type": 1 00:07:05.183 }, 00:07:05.183 { 00:07:05.183 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:05.183 "dma_device_type": 2 00:07:05.183 }, 00:07:05.183 { 00:07:05.183 "dma_device_id": "system", 00:07:05.183 "dma_device_type": 1 00:07:05.183 }, 00:07:05.183 { 00:07:05.183 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:05.183 "dma_device_type": 2 00:07:05.183 } 00:07:05.183 ], 00:07:05.183 "driver_specific": { 00:07:05.183 "raid": { 00:07:05.183 "uuid": "e46bb084-fd0e-49e5-b5c0-3fb0382282b0", 00:07:05.183 "strip_size_kb": 64, 00:07:05.183 "state": "online", 00:07:05.183 "raid_level": "raid0", 00:07:05.183 "superblock": true, 00:07:05.183 "num_base_bdevs": 2, 00:07:05.183 "num_base_bdevs_discovered": 2, 00:07:05.183 "num_base_bdevs_operational": 2, 00:07:05.183 "base_bdevs_list": [ 00:07:05.183 { 00:07:05.183 "name": "pt1", 00:07:05.183 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:05.183 "is_configured": true, 00:07:05.183 "data_offset": 2048, 00:07:05.183 "data_size": 63488 00:07:05.183 }, 00:07:05.183 { 00:07:05.183 "name": "pt2", 00:07:05.183 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:05.183 "is_configured": true, 00:07:05.183 "data_offset": 2048, 00:07:05.183 "data_size": 63488 00:07:05.183 } 00:07:05.183 ] 00:07:05.183 } 00:07:05.183 } 00:07:05.183 }' 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:05.183 pt2' 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.183 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:05.442 [2024-12-07 05:34:38.638685] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' e46bb084-fd0e-49e5-b5c0-3fb0382282b0 '!=' e46bb084-fd0e-49e5-b5c0-3fb0382282b0 ']' 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 72285 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 72285 ']' 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 72285 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72285 00:07:05.442 killing process with pid 72285 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72285' 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 72285 00:07:05.442 [2024-12-07 05:34:38.722903] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:05.442 [2024-12-07 05:34:38.722982] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:05.442 [2024-12-07 05:34:38.723032] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:05.442 [2024-12-07 05:34:38.723041] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:05.442 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 72285 00:07:05.442 [2024-12-07 05:34:38.745338] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:05.702 05:34:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:05.702 00:07:05.702 real 0m3.315s 00:07:05.702 user 0m5.162s 00:07:05.702 sys 0m0.669s 00:07:05.702 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:05.702 05:34:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.702 ************************************ 00:07:05.702 END TEST raid_superblock_test 00:07:05.702 ************************************ 00:07:05.702 05:34:39 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:07:05.702 05:34:39 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:05.702 05:34:39 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:05.702 05:34:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:05.702 ************************************ 00:07:05.702 START TEST raid_read_error_test 00:07:05.702 ************************************ 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 read 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.HD0TAMAOtV 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72486 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72486 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 72486 ']' 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:05.702 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:05.702 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.961 [2024-12-07 05:34:39.123306] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:05.961 [2024-12-07 05:34:39.123511] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72486 ] 00:07:05.961 [2024-12-07 05:34:39.277030] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:05.962 [2024-12-07 05:34:39.301283] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:06.220 [2024-12-07 05:34:39.342996] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:06.220 [2024-12-07 05:34:39.343030] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.790 BaseBdev1_malloc 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.790 true 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.790 [2024-12-07 05:34:39.977903] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:06.790 [2024-12-07 05:34:39.977955] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:06.790 [2024-12-07 05:34:39.977974] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:06.790 [2024-12-07 05:34:39.977989] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:06.790 [2024-12-07 05:34:39.980195] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:06.790 [2024-12-07 05:34:39.980232] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:06.790 BaseBdev1 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.790 BaseBdev2_malloc 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.790 05:34:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.790 true 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.790 [2024-12-07 05:34:40.018400] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:06.790 [2024-12-07 05:34:40.018447] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:06.790 [2024-12-07 05:34:40.018480] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:06.790 [2024-12-07 05:34:40.018497] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:06.790 [2024-12-07 05:34:40.020619] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:06.790 [2024-12-07 05:34:40.020682] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:06.790 BaseBdev2 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.790 [2024-12-07 05:34:40.030418] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:06.790 [2024-12-07 05:34:40.032267] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:06.790 [2024-12-07 05:34:40.032457] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:06.790 [2024-12-07 05:34:40.032470] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:06.790 [2024-12-07 05:34:40.032731] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:06.790 [2024-12-07 05:34:40.032894] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:06.790 [2024-12-07 05:34:40.032907] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:06.790 [2024-12-07 05:34:40.033019] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.790 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:06.790 "name": "raid_bdev1", 00:07:06.790 "uuid": "07d26e9e-75e2-415a-9eeb-b730d59ea611", 00:07:06.791 "strip_size_kb": 64, 00:07:06.791 "state": "online", 00:07:06.791 "raid_level": "raid0", 00:07:06.791 "superblock": true, 00:07:06.791 "num_base_bdevs": 2, 00:07:06.791 "num_base_bdevs_discovered": 2, 00:07:06.791 "num_base_bdevs_operational": 2, 00:07:06.791 "base_bdevs_list": [ 00:07:06.791 { 00:07:06.791 "name": "BaseBdev1", 00:07:06.791 "uuid": "a3ffa41d-b640-5fbd-bacc-567ad1a1d31c", 00:07:06.791 "is_configured": true, 00:07:06.791 "data_offset": 2048, 00:07:06.791 "data_size": 63488 00:07:06.791 }, 00:07:06.791 { 00:07:06.791 "name": "BaseBdev2", 00:07:06.791 "uuid": "a9652d46-885a-5205-be73-b6f7238dfe99", 00:07:06.791 "is_configured": true, 00:07:06.791 "data_offset": 2048, 00:07:06.791 "data_size": 63488 00:07:06.791 } 00:07:06.791 ] 00:07:06.791 }' 00:07:06.791 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:06.791 05:34:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.359 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:07.359 05:34:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:07.359 [2024-12-07 05:34:40.521995] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:08.300 "name": "raid_bdev1", 00:07:08.300 "uuid": "07d26e9e-75e2-415a-9eeb-b730d59ea611", 00:07:08.300 "strip_size_kb": 64, 00:07:08.300 "state": "online", 00:07:08.300 "raid_level": "raid0", 00:07:08.300 "superblock": true, 00:07:08.300 "num_base_bdevs": 2, 00:07:08.300 "num_base_bdevs_discovered": 2, 00:07:08.300 "num_base_bdevs_operational": 2, 00:07:08.300 "base_bdevs_list": [ 00:07:08.300 { 00:07:08.300 "name": "BaseBdev1", 00:07:08.300 "uuid": "a3ffa41d-b640-5fbd-bacc-567ad1a1d31c", 00:07:08.300 "is_configured": true, 00:07:08.300 "data_offset": 2048, 00:07:08.300 "data_size": 63488 00:07:08.300 }, 00:07:08.300 { 00:07:08.300 "name": "BaseBdev2", 00:07:08.300 "uuid": "a9652d46-885a-5205-be73-b6f7238dfe99", 00:07:08.300 "is_configured": true, 00:07:08.300 "data_offset": 2048, 00:07:08.300 "data_size": 63488 00:07:08.300 } 00:07:08.300 ] 00:07:08.300 }' 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:08.300 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.559 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:08.559 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.559 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.559 [2024-12-07 05:34:41.889670] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:08.559 [2024-12-07 05:34:41.889700] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:08.559 [2024-12-07 05:34:41.892286] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:08.559 [2024-12-07 05:34:41.892343] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:08.559 [2024-12-07 05:34:41.892378] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:08.559 [2024-12-07 05:34:41.892392] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:08.559 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.559 05:34:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72486 00:07:08.559 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 72486 ']' 00:07:08.559 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 72486 00:07:08.559 { 00:07:08.559 "results": [ 00:07:08.559 { 00:07:08.559 "job": "raid_bdev1", 00:07:08.559 "core_mask": "0x1", 00:07:08.559 "workload": "randrw", 00:07:08.559 "percentage": 50, 00:07:08.559 "status": "finished", 00:07:08.559 "queue_depth": 1, 00:07:08.559 "io_size": 131072, 00:07:08.559 "runtime": 1.368532, 00:07:08.559 "iops": 17032.11908819085, 00:07:08.559 "mibps": 2129.0148860238564, 00:07:08.559 "io_failed": 1, 00:07:08.559 "io_timeout": 0, 00:07:08.559 "avg_latency_us": 80.90000977896176, 00:07:08.559 "min_latency_us": 25.041048034934498, 00:07:08.559 "max_latency_us": 1438.071615720524 00:07:08.559 } 00:07:08.559 ], 00:07:08.559 "core_count": 1 00:07:08.559 } 00:07:08.559 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:07:08.559 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:08.559 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72486 00:07:08.819 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:08.820 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:08.820 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72486' 00:07:08.820 killing process with pid 72486 00:07:08.820 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 72486 00:07:08.820 [2024-12-07 05:34:41.936909] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:08.820 05:34:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 72486 00:07:08.820 [2024-12-07 05:34:41.951845] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:08.820 05:34:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:08.820 05:34:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.HD0TAMAOtV 00:07:08.820 05:34:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:08.820 05:34:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:07:08.820 05:34:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:08.820 05:34:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:08.820 05:34:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:08.820 05:34:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:07:08.820 00:07:08.820 real 0m3.137s 00:07:08.820 user 0m3.997s 00:07:08.820 sys 0m0.479s 00:07:08.820 05:34:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:08.820 ************************************ 00:07:08.820 END TEST raid_read_error_test 00:07:08.820 ************************************ 00:07:08.820 05:34:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.080 05:34:42 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:07:09.080 05:34:42 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:09.080 05:34:42 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:09.080 05:34:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:09.080 ************************************ 00:07:09.080 START TEST raid_write_error_test 00:07:09.080 ************************************ 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 write 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.KuQQbsbD0O 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72615 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72615 00:07:09.080 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 72615 ']' 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:09.080 05:34:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.080 [2024-12-07 05:34:42.324902] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:09.080 [2024-12-07 05:34:42.325034] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72615 ] 00:07:09.340 [2024-12-07 05:34:42.479015] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:09.340 [2024-12-07 05:34:42.503324] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:09.340 [2024-12-07 05:34:42.544843] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:09.340 [2024-12-07 05:34:42.544892] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.911 BaseBdev1_malloc 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.911 true 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.911 [2024-12-07 05:34:43.175504] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:09.911 [2024-12-07 05:34:43.175558] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:09.911 [2024-12-07 05:34:43.175597] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:09.911 [2024-12-07 05:34:43.175608] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:09.911 [2024-12-07 05:34:43.177749] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:09.911 [2024-12-07 05:34:43.177784] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:09.911 BaseBdev1 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.911 BaseBdev2_malloc 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.911 true 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.911 [2024-12-07 05:34:43.215834] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:09.911 [2024-12-07 05:34:43.215881] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:09.911 [2024-12-07 05:34:43.215915] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:09.911 [2024-12-07 05:34:43.215943] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:09.911 [2024-12-07 05:34:43.217979] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:09.911 [2024-12-07 05:34:43.218016] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:09.911 BaseBdev2 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.911 [2024-12-07 05:34:43.227860] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:09.911 [2024-12-07 05:34:43.229680] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:09.911 [2024-12-07 05:34:43.229850] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:09.911 [2024-12-07 05:34:43.229862] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:09.911 [2024-12-07 05:34:43.230116] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:09.911 [2024-12-07 05:34:43.230268] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:09.911 [2024-12-07 05:34:43.230280] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:09.911 [2024-12-07 05:34:43.230407] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:09.911 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:09.912 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:09.912 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:09.912 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:09.912 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.912 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.912 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:09.912 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:10.174 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:10.174 "name": "raid_bdev1", 00:07:10.174 "uuid": "272bf971-3d33-4afa-a532-c217881d1d33", 00:07:10.174 "strip_size_kb": 64, 00:07:10.174 "state": "online", 00:07:10.174 "raid_level": "raid0", 00:07:10.174 "superblock": true, 00:07:10.174 "num_base_bdevs": 2, 00:07:10.174 "num_base_bdevs_discovered": 2, 00:07:10.174 "num_base_bdevs_operational": 2, 00:07:10.174 "base_bdevs_list": [ 00:07:10.174 { 00:07:10.174 "name": "BaseBdev1", 00:07:10.174 "uuid": "554fcacc-2685-5247-b7fc-3886b216069e", 00:07:10.174 "is_configured": true, 00:07:10.174 "data_offset": 2048, 00:07:10.174 "data_size": 63488 00:07:10.174 }, 00:07:10.174 { 00:07:10.174 "name": "BaseBdev2", 00:07:10.174 "uuid": "eb94b1a3-fac1-526b-8f97-0e4fe4f326cd", 00:07:10.174 "is_configured": true, 00:07:10.174 "data_offset": 2048, 00:07:10.174 "data_size": 63488 00:07:10.174 } 00:07:10.174 ] 00:07:10.174 }' 00:07:10.174 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:10.174 05:34:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.434 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:10.434 05:34:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:10.434 [2024-12-07 05:34:43.771266] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.374 05:34:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.634 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:11.634 "name": "raid_bdev1", 00:07:11.634 "uuid": "272bf971-3d33-4afa-a532-c217881d1d33", 00:07:11.634 "strip_size_kb": 64, 00:07:11.634 "state": "online", 00:07:11.634 "raid_level": "raid0", 00:07:11.634 "superblock": true, 00:07:11.634 "num_base_bdevs": 2, 00:07:11.634 "num_base_bdevs_discovered": 2, 00:07:11.634 "num_base_bdevs_operational": 2, 00:07:11.634 "base_bdevs_list": [ 00:07:11.634 { 00:07:11.634 "name": "BaseBdev1", 00:07:11.634 "uuid": "554fcacc-2685-5247-b7fc-3886b216069e", 00:07:11.634 "is_configured": true, 00:07:11.634 "data_offset": 2048, 00:07:11.634 "data_size": 63488 00:07:11.634 }, 00:07:11.634 { 00:07:11.635 "name": "BaseBdev2", 00:07:11.635 "uuid": "eb94b1a3-fac1-526b-8f97-0e4fe4f326cd", 00:07:11.635 "is_configured": true, 00:07:11.635 "data_offset": 2048, 00:07:11.635 "data_size": 63488 00:07:11.635 } 00:07:11.635 ] 00:07:11.635 }' 00:07:11.635 05:34:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:11.635 05:34:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.895 [2024-12-07 05:34:45.163064] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:11.895 [2024-12-07 05:34:45.163161] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:11.895 [2024-12-07 05:34:45.165789] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:11.895 [2024-12-07 05:34:45.165869] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:11.895 [2024-12-07 05:34:45.165922] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:11.895 [2024-12-07 05:34:45.165962] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:11.895 { 00:07:11.895 "results": [ 00:07:11.895 { 00:07:11.895 "job": "raid_bdev1", 00:07:11.895 "core_mask": "0x1", 00:07:11.895 "workload": "randrw", 00:07:11.895 "percentage": 50, 00:07:11.895 "status": "finished", 00:07:11.895 "queue_depth": 1, 00:07:11.895 "io_size": 131072, 00:07:11.895 "runtime": 1.392868, 00:07:11.895 "iops": 17130.122883144704, 00:07:11.895 "mibps": 2141.265360393088, 00:07:11.895 "io_failed": 1, 00:07:11.895 "io_timeout": 0, 00:07:11.895 "avg_latency_us": 80.4855035779457, 00:07:11.895 "min_latency_us": 25.152838427947597, 00:07:11.895 "max_latency_us": 1387.989519650655 00:07:11.895 } 00:07:11.895 ], 00:07:11.895 "core_count": 1 00:07:11.895 } 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72615 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 72615 ']' 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 72615 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72615 00:07:11.895 killing process with pid 72615 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72615' 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 72615 00:07:11.895 [2024-12-07 05:34:45.210038] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:11.895 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 72615 00:07:11.895 [2024-12-07 05:34:45.225635] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:12.155 05:34:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.KuQQbsbD0O 00:07:12.155 05:34:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:12.155 05:34:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:12.155 05:34:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:07:12.155 05:34:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:12.155 ************************************ 00:07:12.155 END TEST raid_write_error_test 00:07:12.155 ************************************ 00:07:12.155 05:34:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:12.155 05:34:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:12.155 05:34:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:07:12.155 00:07:12.155 real 0m3.206s 00:07:12.155 user 0m4.128s 00:07:12.155 sys 0m0.485s 00:07:12.155 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:12.155 05:34:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.155 05:34:45 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:12.155 05:34:45 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:07:12.155 05:34:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:12.155 05:34:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:12.155 05:34:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:12.155 ************************************ 00:07:12.155 START TEST raid_state_function_test 00:07:12.155 ************************************ 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 false 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=72742 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72742' 00:07:12.155 Process raid pid: 72742 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 72742 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 72742 ']' 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:12.155 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:12.155 05:34:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.414 [2024-12-07 05:34:45.589584] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:12.414 [2024-12-07 05:34:45.589740] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:12.414 [2024-12-07 05:34:45.744406] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:12.414 [2024-12-07 05:34:45.768947] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:12.673 [2024-12-07 05:34:45.810064] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:12.673 [2024-12-07 05:34:45.810100] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.242 [2024-12-07 05:34:46.416042] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:13.242 [2024-12-07 05:34:46.416102] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:13.242 [2024-12-07 05:34:46.416112] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:13.242 [2024-12-07 05:34:46.416122] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.242 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.243 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:13.243 "name": "Existed_Raid", 00:07:13.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:13.243 "strip_size_kb": 64, 00:07:13.243 "state": "configuring", 00:07:13.243 "raid_level": "concat", 00:07:13.243 "superblock": false, 00:07:13.243 "num_base_bdevs": 2, 00:07:13.243 "num_base_bdevs_discovered": 0, 00:07:13.243 "num_base_bdevs_operational": 2, 00:07:13.243 "base_bdevs_list": [ 00:07:13.243 { 00:07:13.243 "name": "BaseBdev1", 00:07:13.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:13.243 "is_configured": false, 00:07:13.243 "data_offset": 0, 00:07:13.243 "data_size": 0 00:07:13.243 }, 00:07:13.243 { 00:07:13.243 "name": "BaseBdev2", 00:07:13.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:13.243 "is_configured": false, 00:07:13.243 "data_offset": 0, 00:07:13.243 "data_size": 0 00:07:13.243 } 00:07:13.243 ] 00:07:13.243 }' 00:07:13.243 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:13.243 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.513 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:13.513 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.513 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.513 [2024-12-07 05:34:46.847222] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:13.513 [2024-12-07 05:34:46.847328] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:13.513 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.513 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:13.513 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.513 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.513 [2024-12-07 05:34:46.859213] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:13.513 [2024-12-07 05:34:46.859257] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:13.513 [2024-12-07 05:34:46.859265] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:13.513 [2024-12-07 05:34:46.859285] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:13.513 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.513 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:13.513 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.513 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.795 [2024-12-07 05:34:46.879833] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:13.795 BaseBdev1 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.795 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.795 [ 00:07:13.795 { 00:07:13.795 "name": "BaseBdev1", 00:07:13.795 "aliases": [ 00:07:13.795 "5d79f733-1290-4def-8322-62437faf3c16" 00:07:13.795 ], 00:07:13.795 "product_name": "Malloc disk", 00:07:13.795 "block_size": 512, 00:07:13.795 "num_blocks": 65536, 00:07:13.795 "uuid": "5d79f733-1290-4def-8322-62437faf3c16", 00:07:13.795 "assigned_rate_limits": { 00:07:13.795 "rw_ios_per_sec": 0, 00:07:13.795 "rw_mbytes_per_sec": 0, 00:07:13.795 "r_mbytes_per_sec": 0, 00:07:13.795 "w_mbytes_per_sec": 0 00:07:13.795 }, 00:07:13.795 "claimed": true, 00:07:13.795 "claim_type": "exclusive_write", 00:07:13.796 "zoned": false, 00:07:13.796 "supported_io_types": { 00:07:13.796 "read": true, 00:07:13.796 "write": true, 00:07:13.796 "unmap": true, 00:07:13.796 "flush": true, 00:07:13.796 "reset": true, 00:07:13.796 "nvme_admin": false, 00:07:13.796 "nvme_io": false, 00:07:13.796 "nvme_io_md": false, 00:07:13.796 "write_zeroes": true, 00:07:13.796 "zcopy": true, 00:07:13.796 "get_zone_info": false, 00:07:13.796 "zone_management": false, 00:07:13.796 "zone_append": false, 00:07:13.796 "compare": false, 00:07:13.796 "compare_and_write": false, 00:07:13.796 "abort": true, 00:07:13.796 "seek_hole": false, 00:07:13.796 "seek_data": false, 00:07:13.796 "copy": true, 00:07:13.796 "nvme_iov_md": false 00:07:13.796 }, 00:07:13.796 "memory_domains": [ 00:07:13.796 { 00:07:13.796 "dma_device_id": "system", 00:07:13.796 "dma_device_type": 1 00:07:13.796 }, 00:07:13.796 { 00:07:13.796 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:13.796 "dma_device_type": 2 00:07:13.796 } 00:07:13.796 ], 00:07:13.796 "driver_specific": {} 00:07:13.796 } 00:07:13.796 ] 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:13.796 "name": "Existed_Raid", 00:07:13.796 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:13.796 "strip_size_kb": 64, 00:07:13.796 "state": "configuring", 00:07:13.796 "raid_level": "concat", 00:07:13.796 "superblock": false, 00:07:13.796 "num_base_bdevs": 2, 00:07:13.796 "num_base_bdevs_discovered": 1, 00:07:13.796 "num_base_bdevs_operational": 2, 00:07:13.796 "base_bdevs_list": [ 00:07:13.796 { 00:07:13.796 "name": "BaseBdev1", 00:07:13.796 "uuid": "5d79f733-1290-4def-8322-62437faf3c16", 00:07:13.796 "is_configured": true, 00:07:13.796 "data_offset": 0, 00:07:13.796 "data_size": 65536 00:07:13.796 }, 00:07:13.796 { 00:07:13.796 "name": "BaseBdev2", 00:07:13.796 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:13.796 "is_configured": false, 00:07:13.796 "data_offset": 0, 00:07:13.796 "data_size": 0 00:07:13.796 } 00:07:13.796 ] 00:07:13.796 }' 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:13.796 05:34:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.057 [2024-12-07 05:34:47.371032] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:14.057 [2024-12-07 05:34:47.371074] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.057 [2024-12-07 05:34:47.383047] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:14.057 [2024-12-07 05:34:47.384942] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:14.057 [2024-12-07 05:34:47.384981] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:14.057 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.317 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:14.317 "name": "Existed_Raid", 00:07:14.317 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:14.317 "strip_size_kb": 64, 00:07:14.317 "state": "configuring", 00:07:14.317 "raid_level": "concat", 00:07:14.317 "superblock": false, 00:07:14.317 "num_base_bdevs": 2, 00:07:14.317 "num_base_bdevs_discovered": 1, 00:07:14.317 "num_base_bdevs_operational": 2, 00:07:14.317 "base_bdevs_list": [ 00:07:14.317 { 00:07:14.317 "name": "BaseBdev1", 00:07:14.317 "uuid": "5d79f733-1290-4def-8322-62437faf3c16", 00:07:14.317 "is_configured": true, 00:07:14.317 "data_offset": 0, 00:07:14.317 "data_size": 65536 00:07:14.317 }, 00:07:14.317 { 00:07:14.317 "name": "BaseBdev2", 00:07:14.317 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:14.317 "is_configured": false, 00:07:14.317 "data_offset": 0, 00:07:14.317 "data_size": 0 00:07:14.317 } 00:07:14.317 ] 00:07:14.317 }' 00:07:14.317 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:14.317 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.577 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:14.577 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.577 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.577 [2024-12-07 05:34:47.841218] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:14.577 [2024-12-07 05:34:47.841322] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:14.578 [2024-12-07 05:34:47.841348] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:14.578 [2024-12-07 05:34:47.841667] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:14.578 [2024-12-07 05:34:47.841864] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:14.578 [2024-12-07 05:34:47.841912] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:14.578 [2024-12-07 05:34:47.842132] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:14.578 BaseBdev2 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.578 [ 00:07:14.578 { 00:07:14.578 "name": "BaseBdev2", 00:07:14.578 "aliases": [ 00:07:14.578 "3027ecc5-d6d5-4311-bcdc-76d7a8d71d7c" 00:07:14.578 ], 00:07:14.578 "product_name": "Malloc disk", 00:07:14.578 "block_size": 512, 00:07:14.578 "num_blocks": 65536, 00:07:14.578 "uuid": "3027ecc5-d6d5-4311-bcdc-76d7a8d71d7c", 00:07:14.578 "assigned_rate_limits": { 00:07:14.578 "rw_ios_per_sec": 0, 00:07:14.578 "rw_mbytes_per_sec": 0, 00:07:14.578 "r_mbytes_per_sec": 0, 00:07:14.578 "w_mbytes_per_sec": 0 00:07:14.578 }, 00:07:14.578 "claimed": true, 00:07:14.578 "claim_type": "exclusive_write", 00:07:14.578 "zoned": false, 00:07:14.578 "supported_io_types": { 00:07:14.578 "read": true, 00:07:14.578 "write": true, 00:07:14.578 "unmap": true, 00:07:14.578 "flush": true, 00:07:14.578 "reset": true, 00:07:14.578 "nvme_admin": false, 00:07:14.578 "nvme_io": false, 00:07:14.578 "nvme_io_md": false, 00:07:14.578 "write_zeroes": true, 00:07:14.578 "zcopy": true, 00:07:14.578 "get_zone_info": false, 00:07:14.578 "zone_management": false, 00:07:14.578 "zone_append": false, 00:07:14.578 "compare": false, 00:07:14.578 "compare_and_write": false, 00:07:14.578 "abort": true, 00:07:14.578 "seek_hole": false, 00:07:14.578 "seek_data": false, 00:07:14.578 "copy": true, 00:07:14.578 "nvme_iov_md": false 00:07:14.578 }, 00:07:14.578 "memory_domains": [ 00:07:14.578 { 00:07:14.578 "dma_device_id": "system", 00:07:14.578 "dma_device_type": 1 00:07:14.578 }, 00:07:14.578 { 00:07:14.578 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:14.578 "dma_device_type": 2 00:07:14.578 } 00:07:14.578 ], 00:07:14.578 "driver_specific": {} 00:07:14.578 } 00:07:14.578 ] 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:14.578 "name": "Existed_Raid", 00:07:14.578 "uuid": "b2ad7762-23da-43d4-bb15-b8fac1fa9f7e", 00:07:14.578 "strip_size_kb": 64, 00:07:14.578 "state": "online", 00:07:14.578 "raid_level": "concat", 00:07:14.578 "superblock": false, 00:07:14.578 "num_base_bdevs": 2, 00:07:14.578 "num_base_bdevs_discovered": 2, 00:07:14.578 "num_base_bdevs_operational": 2, 00:07:14.578 "base_bdevs_list": [ 00:07:14.578 { 00:07:14.578 "name": "BaseBdev1", 00:07:14.578 "uuid": "5d79f733-1290-4def-8322-62437faf3c16", 00:07:14.578 "is_configured": true, 00:07:14.578 "data_offset": 0, 00:07:14.578 "data_size": 65536 00:07:14.578 }, 00:07:14.578 { 00:07:14.578 "name": "BaseBdev2", 00:07:14.578 "uuid": "3027ecc5-d6d5-4311-bcdc-76d7a8d71d7c", 00:07:14.578 "is_configured": true, 00:07:14.578 "data_offset": 0, 00:07:14.578 "data_size": 65536 00:07:14.578 } 00:07:14.578 ] 00:07:14.578 }' 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:14.578 05:34:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.147 [2024-12-07 05:34:48.308727] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.147 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:15.147 "name": "Existed_Raid", 00:07:15.147 "aliases": [ 00:07:15.147 "b2ad7762-23da-43d4-bb15-b8fac1fa9f7e" 00:07:15.147 ], 00:07:15.147 "product_name": "Raid Volume", 00:07:15.147 "block_size": 512, 00:07:15.147 "num_blocks": 131072, 00:07:15.147 "uuid": "b2ad7762-23da-43d4-bb15-b8fac1fa9f7e", 00:07:15.147 "assigned_rate_limits": { 00:07:15.147 "rw_ios_per_sec": 0, 00:07:15.147 "rw_mbytes_per_sec": 0, 00:07:15.147 "r_mbytes_per_sec": 0, 00:07:15.147 "w_mbytes_per_sec": 0 00:07:15.147 }, 00:07:15.147 "claimed": false, 00:07:15.147 "zoned": false, 00:07:15.147 "supported_io_types": { 00:07:15.147 "read": true, 00:07:15.147 "write": true, 00:07:15.147 "unmap": true, 00:07:15.147 "flush": true, 00:07:15.147 "reset": true, 00:07:15.147 "nvme_admin": false, 00:07:15.147 "nvme_io": false, 00:07:15.147 "nvme_io_md": false, 00:07:15.147 "write_zeroes": true, 00:07:15.147 "zcopy": false, 00:07:15.147 "get_zone_info": false, 00:07:15.147 "zone_management": false, 00:07:15.147 "zone_append": false, 00:07:15.147 "compare": false, 00:07:15.147 "compare_and_write": false, 00:07:15.147 "abort": false, 00:07:15.147 "seek_hole": false, 00:07:15.147 "seek_data": false, 00:07:15.147 "copy": false, 00:07:15.147 "nvme_iov_md": false 00:07:15.147 }, 00:07:15.147 "memory_domains": [ 00:07:15.147 { 00:07:15.147 "dma_device_id": "system", 00:07:15.147 "dma_device_type": 1 00:07:15.147 }, 00:07:15.147 { 00:07:15.147 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:15.148 "dma_device_type": 2 00:07:15.148 }, 00:07:15.148 { 00:07:15.148 "dma_device_id": "system", 00:07:15.148 "dma_device_type": 1 00:07:15.148 }, 00:07:15.148 { 00:07:15.148 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:15.148 "dma_device_type": 2 00:07:15.148 } 00:07:15.148 ], 00:07:15.148 "driver_specific": { 00:07:15.148 "raid": { 00:07:15.148 "uuid": "b2ad7762-23da-43d4-bb15-b8fac1fa9f7e", 00:07:15.148 "strip_size_kb": 64, 00:07:15.148 "state": "online", 00:07:15.148 "raid_level": "concat", 00:07:15.148 "superblock": false, 00:07:15.148 "num_base_bdevs": 2, 00:07:15.148 "num_base_bdevs_discovered": 2, 00:07:15.148 "num_base_bdevs_operational": 2, 00:07:15.148 "base_bdevs_list": [ 00:07:15.148 { 00:07:15.148 "name": "BaseBdev1", 00:07:15.148 "uuid": "5d79f733-1290-4def-8322-62437faf3c16", 00:07:15.148 "is_configured": true, 00:07:15.148 "data_offset": 0, 00:07:15.148 "data_size": 65536 00:07:15.148 }, 00:07:15.148 { 00:07:15.148 "name": "BaseBdev2", 00:07:15.148 "uuid": "3027ecc5-d6d5-4311-bcdc-76d7a8d71d7c", 00:07:15.148 "is_configured": true, 00:07:15.148 "data_offset": 0, 00:07:15.148 "data_size": 65536 00:07:15.148 } 00:07:15.148 ] 00:07:15.148 } 00:07:15.148 } 00:07:15.148 }' 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:15.148 BaseBdev2' 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:15.148 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.408 [2024-12-07 05:34:48.528118] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:15.408 [2024-12-07 05:34:48.528151] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:15.408 [2024-12-07 05:34:48.528215] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.408 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:15.408 "name": "Existed_Raid", 00:07:15.408 "uuid": "b2ad7762-23da-43d4-bb15-b8fac1fa9f7e", 00:07:15.408 "strip_size_kb": 64, 00:07:15.408 "state": "offline", 00:07:15.408 "raid_level": "concat", 00:07:15.408 "superblock": false, 00:07:15.408 "num_base_bdevs": 2, 00:07:15.408 "num_base_bdevs_discovered": 1, 00:07:15.408 "num_base_bdevs_operational": 1, 00:07:15.408 "base_bdevs_list": [ 00:07:15.408 { 00:07:15.408 "name": null, 00:07:15.409 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:15.409 "is_configured": false, 00:07:15.409 "data_offset": 0, 00:07:15.409 "data_size": 65536 00:07:15.409 }, 00:07:15.409 { 00:07:15.409 "name": "BaseBdev2", 00:07:15.409 "uuid": "3027ecc5-d6d5-4311-bcdc-76d7a8d71d7c", 00:07:15.409 "is_configured": true, 00:07:15.409 "data_offset": 0, 00:07:15.409 "data_size": 65536 00:07:15.409 } 00:07:15.409 ] 00:07:15.409 }' 00:07:15.409 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:15.409 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.669 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:15.669 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:15.669 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:15.669 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.669 05:34:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:15.669 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.669 05:34:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.669 [2024-12-07 05:34:49.010511] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:15.669 [2024-12-07 05:34:49.010605] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.669 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 72742 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 72742 ']' 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 72742 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72742 00:07:15.929 killing process with pid 72742 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72742' 00:07:15.929 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 72742 00:07:15.930 [2024-12-07 05:34:49.104469] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:15.930 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 72742 00:07:15.930 [2024-12-07 05:34:49.105420] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:16.190 05:34:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:16.190 00:07:16.190 real 0m3.812s 00:07:16.190 user 0m6.080s 00:07:16.190 sys 0m0.697s 00:07:16.190 ************************************ 00:07:16.190 END TEST raid_state_function_test 00:07:16.190 ************************************ 00:07:16.190 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:16.190 05:34:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.190 05:34:49 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:07:16.190 05:34:49 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:16.190 05:34:49 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:16.190 05:34:49 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:16.190 ************************************ 00:07:16.190 START TEST raid_state_function_test_sb 00:07:16.190 ************************************ 00:07:16.190 05:34:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 true 00:07:16.190 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:16.190 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:16.190 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:16.190 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:16.190 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:16.190 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72984 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72984' 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:16.191 Process raid pid: 72984 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72984 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 72984 ']' 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:16.191 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:16.191 05:34:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:16.191 [2024-12-07 05:34:49.474096] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:16.191 [2024-12-07 05:34:49.474328] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:16.451 [2024-12-07 05:34:49.606497] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:16.451 [2024-12-07 05:34:49.630552] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:16.451 [2024-12-07 05:34:49.672025] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:16.451 [2024-12-07 05:34:49.672143] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.020 [2024-12-07 05:34:50.306029] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:17.020 [2024-12-07 05:34:50.306140] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:17.020 [2024-12-07 05:34:50.306170] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:17.020 [2024-12-07 05:34:50.306193] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:17.020 "name": "Existed_Raid", 00:07:17.020 "uuid": "3c8354fb-d2ed-4bf2-85ef-1b201dd3c7ed", 00:07:17.020 "strip_size_kb": 64, 00:07:17.020 "state": "configuring", 00:07:17.020 "raid_level": "concat", 00:07:17.020 "superblock": true, 00:07:17.020 "num_base_bdevs": 2, 00:07:17.020 "num_base_bdevs_discovered": 0, 00:07:17.020 "num_base_bdevs_operational": 2, 00:07:17.020 "base_bdevs_list": [ 00:07:17.020 { 00:07:17.020 "name": "BaseBdev1", 00:07:17.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:17.020 "is_configured": false, 00:07:17.020 "data_offset": 0, 00:07:17.020 "data_size": 0 00:07:17.020 }, 00:07:17.020 { 00:07:17.020 "name": "BaseBdev2", 00:07:17.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:17.020 "is_configured": false, 00:07:17.020 "data_offset": 0, 00:07:17.020 "data_size": 0 00:07:17.020 } 00:07:17.020 ] 00:07:17.020 }' 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:17.020 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.590 [2024-12-07 05:34:50.689281] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:17.590 [2024-12-07 05:34:50.689324] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.590 [2024-12-07 05:34:50.701282] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:17.590 [2024-12-07 05:34:50.701364] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:17.590 [2024-12-07 05:34:50.701410] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:17.590 [2024-12-07 05:34:50.701448] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.590 [2024-12-07 05:34:50.721962] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:17.590 BaseBdev1 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.590 [ 00:07:17.590 { 00:07:17.590 "name": "BaseBdev1", 00:07:17.590 "aliases": [ 00:07:17.590 "2c766a7d-2184-44a9-8e67-8a028d662daa" 00:07:17.590 ], 00:07:17.590 "product_name": "Malloc disk", 00:07:17.590 "block_size": 512, 00:07:17.590 "num_blocks": 65536, 00:07:17.590 "uuid": "2c766a7d-2184-44a9-8e67-8a028d662daa", 00:07:17.590 "assigned_rate_limits": { 00:07:17.590 "rw_ios_per_sec": 0, 00:07:17.590 "rw_mbytes_per_sec": 0, 00:07:17.590 "r_mbytes_per_sec": 0, 00:07:17.590 "w_mbytes_per_sec": 0 00:07:17.590 }, 00:07:17.590 "claimed": true, 00:07:17.590 "claim_type": "exclusive_write", 00:07:17.590 "zoned": false, 00:07:17.590 "supported_io_types": { 00:07:17.590 "read": true, 00:07:17.590 "write": true, 00:07:17.590 "unmap": true, 00:07:17.590 "flush": true, 00:07:17.590 "reset": true, 00:07:17.590 "nvme_admin": false, 00:07:17.590 "nvme_io": false, 00:07:17.590 "nvme_io_md": false, 00:07:17.590 "write_zeroes": true, 00:07:17.590 "zcopy": true, 00:07:17.590 "get_zone_info": false, 00:07:17.590 "zone_management": false, 00:07:17.590 "zone_append": false, 00:07:17.590 "compare": false, 00:07:17.590 "compare_and_write": false, 00:07:17.590 "abort": true, 00:07:17.590 "seek_hole": false, 00:07:17.590 "seek_data": false, 00:07:17.590 "copy": true, 00:07:17.590 "nvme_iov_md": false 00:07:17.590 }, 00:07:17.590 "memory_domains": [ 00:07:17.590 { 00:07:17.590 "dma_device_id": "system", 00:07:17.590 "dma_device_type": 1 00:07:17.590 }, 00:07:17.590 { 00:07:17.590 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:17.590 "dma_device_type": 2 00:07:17.590 } 00:07:17.590 ], 00:07:17.590 "driver_specific": {} 00:07:17.590 } 00:07:17.590 ] 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:17.590 "name": "Existed_Raid", 00:07:17.590 "uuid": "7f775011-2d54-48d2-ab57-7e0226f285bb", 00:07:17.590 "strip_size_kb": 64, 00:07:17.590 "state": "configuring", 00:07:17.590 "raid_level": "concat", 00:07:17.590 "superblock": true, 00:07:17.590 "num_base_bdevs": 2, 00:07:17.590 "num_base_bdevs_discovered": 1, 00:07:17.590 "num_base_bdevs_operational": 2, 00:07:17.590 "base_bdevs_list": [ 00:07:17.590 { 00:07:17.590 "name": "BaseBdev1", 00:07:17.590 "uuid": "2c766a7d-2184-44a9-8e67-8a028d662daa", 00:07:17.590 "is_configured": true, 00:07:17.590 "data_offset": 2048, 00:07:17.590 "data_size": 63488 00:07:17.590 }, 00:07:17.590 { 00:07:17.590 "name": "BaseBdev2", 00:07:17.590 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:17.590 "is_configured": false, 00:07:17.590 "data_offset": 0, 00:07:17.590 "data_size": 0 00:07:17.590 } 00:07:17.590 ] 00:07:17.590 }' 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:17.590 05:34:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.850 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:17.850 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.850 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.850 [2024-12-07 05:34:51.137282] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:17.850 [2024-12-07 05:34:51.137328] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:17.850 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.850 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:17.850 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.851 [2024-12-07 05:34:51.149293] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:17.851 [2024-12-07 05:34:51.151111] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:17.851 [2024-12-07 05:34:51.151152] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:17.851 "name": "Existed_Raid", 00:07:17.851 "uuid": "6572ec0c-2ad3-4da6-aeb9-1f9b21ccc748", 00:07:17.851 "strip_size_kb": 64, 00:07:17.851 "state": "configuring", 00:07:17.851 "raid_level": "concat", 00:07:17.851 "superblock": true, 00:07:17.851 "num_base_bdevs": 2, 00:07:17.851 "num_base_bdevs_discovered": 1, 00:07:17.851 "num_base_bdevs_operational": 2, 00:07:17.851 "base_bdevs_list": [ 00:07:17.851 { 00:07:17.851 "name": "BaseBdev1", 00:07:17.851 "uuid": "2c766a7d-2184-44a9-8e67-8a028d662daa", 00:07:17.851 "is_configured": true, 00:07:17.851 "data_offset": 2048, 00:07:17.851 "data_size": 63488 00:07:17.851 }, 00:07:17.851 { 00:07:17.851 "name": "BaseBdev2", 00:07:17.851 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:17.851 "is_configured": false, 00:07:17.851 "data_offset": 0, 00:07:17.851 "data_size": 0 00:07:17.851 } 00:07:17.851 ] 00:07:17.851 }' 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:17.851 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:18.419 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:18.419 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.419 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:18.419 [2024-12-07 05:34:51.591375] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:18.419 [2024-12-07 05:34:51.591658] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:18.419 [2024-12-07 05:34:51.591699] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:18.419 BaseBdev2 00:07:18.419 [2024-12-07 05:34:51.592015] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:18.419 [2024-12-07 05:34:51.592152] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:18.420 [2024-12-07 05:34:51.592214] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:18.420 [2024-12-07 05:34:51.592373] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:18.420 [ 00:07:18.420 { 00:07:18.420 "name": "BaseBdev2", 00:07:18.420 "aliases": [ 00:07:18.420 "6ce2ae51-6b43-44b5-b798-bbde28160712" 00:07:18.420 ], 00:07:18.420 "product_name": "Malloc disk", 00:07:18.420 "block_size": 512, 00:07:18.420 "num_blocks": 65536, 00:07:18.420 "uuid": "6ce2ae51-6b43-44b5-b798-bbde28160712", 00:07:18.420 "assigned_rate_limits": { 00:07:18.420 "rw_ios_per_sec": 0, 00:07:18.420 "rw_mbytes_per_sec": 0, 00:07:18.420 "r_mbytes_per_sec": 0, 00:07:18.420 "w_mbytes_per_sec": 0 00:07:18.420 }, 00:07:18.420 "claimed": true, 00:07:18.420 "claim_type": "exclusive_write", 00:07:18.420 "zoned": false, 00:07:18.420 "supported_io_types": { 00:07:18.420 "read": true, 00:07:18.420 "write": true, 00:07:18.420 "unmap": true, 00:07:18.420 "flush": true, 00:07:18.420 "reset": true, 00:07:18.420 "nvme_admin": false, 00:07:18.420 "nvme_io": false, 00:07:18.420 "nvme_io_md": false, 00:07:18.420 "write_zeroes": true, 00:07:18.420 "zcopy": true, 00:07:18.420 "get_zone_info": false, 00:07:18.420 "zone_management": false, 00:07:18.420 "zone_append": false, 00:07:18.420 "compare": false, 00:07:18.420 "compare_and_write": false, 00:07:18.420 "abort": true, 00:07:18.420 "seek_hole": false, 00:07:18.420 "seek_data": false, 00:07:18.420 "copy": true, 00:07:18.420 "nvme_iov_md": false 00:07:18.420 }, 00:07:18.420 "memory_domains": [ 00:07:18.420 { 00:07:18.420 "dma_device_id": "system", 00:07:18.420 "dma_device_type": 1 00:07:18.420 }, 00:07:18.420 { 00:07:18.420 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:18.420 "dma_device_type": 2 00:07:18.420 } 00:07:18.420 ], 00:07:18.420 "driver_specific": {} 00:07:18.420 } 00:07:18.420 ] 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:18.420 "name": "Existed_Raid", 00:07:18.420 "uuid": "6572ec0c-2ad3-4da6-aeb9-1f9b21ccc748", 00:07:18.420 "strip_size_kb": 64, 00:07:18.420 "state": "online", 00:07:18.420 "raid_level": "concat", 00:07:18.420 "superblock": true, 00:07:18.420 "num_base_bdevs": 2, 00:07:18.420 "num_base_bdevs_discovered": 2, 00:07:18.420 "num_base_bdevs_operational": 2, 00:07:18.420 "base_bdevs_list": [ 00:07:18.420 { 00:07:18.420 "name": "BaseBdev1", 00:07:18.420 "uuid": "2c766a7d-2184-44a9-8e67-8a028d662daa", 00:07:18.420 "is_configured": true, 00:07:18.420 "data_offset": 2048, 00:07:18.420 "data_size": 63488 00:07:18.420 }, 00:07:18.420 { 00:07:18.420 "name": "BaseBdev2", 00:07:18.420 "uuid": "6ce2ae51-6b43-44b5-b798-bbde28160712", 00:07:18.420 "is_configured": true, 00:07:18.420 "data_offset": 2048, 00:07:18.420 "data_size": 63488 00:07:18.420 } 00:07:18.420 ] 00:07:18.420 }' 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:18.420 05:34:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:19.002 [2024-12-07 05:34:52.066903] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.002 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:19.002 "name": "Existed_Raid", 00:07:19.002 "aliases": [ 00:07:19.002 "6572ec0c-2ad3-4da6-aeb9-1f9b21ccc748" 00:07:19.002 ], 00:07:19.002 "product_name": "Raid Volume", 00:07:19.002 "block_size": 512, 00:07:19.002 "num_blocks": 126976, 00:07:19.002 "uuid": "6572ec0c-2ad3-4da6-aeb9-1f9b21ccc748", 00:07:19.002 "assigned_rate_limits": { 00:07:19.002 "rw_ios_per_sec": 0, 00:07:19.002 "rw_mbytes_per_sec": 0, 00:07:19.002 "r_mbytes_per_sec": 0, 00:07:19.002 "w_mbytes_per_sec": 0 00:07:19.002 }, 00:07:19.002 "claimed": false, 00:07:19.002 "zoned": false, 00:07:19.002 "supported_io_types": { 00:07:19.002 "read": true, 00:07:19.002 "write": true, 00:07:19.002 "unmap": true, 00:07:19.002 "flush": true, 00:07:19.002 "reset": true, 00:07:19.002 "nvme_admin": false, 00:07:19.002 "nvme_io": false, 00:07:19.002 "nvme_io_md": false, 00:07:19.002 "write_zeroes": true, 00:07:19.002 "zcopy": false, 00:07:19.002 "get_zone_info": false, 00:07:19.002 "zone_management": false, 00:07:19.002 "zone_append": false, 00:07:19.002 "compare": false, 00:07:19.002 "compare_and_write": false, 00:07:19.002 "abort": false, 00:07:19.002 "seek_hole": false, 00:07:19.002 "seek_data": false, 00:07:19.002 "copy": false, 00:07:19.002 "nvme_iov_md": false 00:07:19.002 }, 00:07:19.002 "memory_domains": [ 00:07:19.002 { 00:07:19.002 "dma_device_id": "system", 00:07:19.002 "dma_device_type": 1 00:07:19.002 }, 00:07:19.002 { 00:07:19.002 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:19.002 "dma_device_type": 2 00:07:19.002 }, 00:07:19.002 { 00:07:19.002 "dma_device_id": "system", 00:07:19.002 "dma_device_type": 1 00:07:19.002 }, 00:07:19.002 { 00:07:19.002 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:19.002 "dma_device_type": 2 00:07:19.002 } 00:07:19.002 ], 00:07:19.002 "driver_specific": { 00:07:19.002 "raid": { 00:07:19.002 "uuid": "6572ec0c-2ad3-4da6-aeb9-1f9b21ccc748", 00:07:19.002 "strip_size_kb": 64, 00:07:19.002 "state": "online", 00:07:19.002 "raid_level": "concat", 00:07:19.002 "superblock": true, 00:07:19.002 "num_base_bdevs": 2, 00:07:19.002 "num_base_bdevs_discovered": 2, 00:07:19.002 "num_base_bdevs_operational": 2, 00:07:19.002 "base_bdevs_list": [ 00:07:19.002 { 00:07:19.002 "name": "BaseBdev1", 00:07:19.002 "uuid": "2c766a7d-2184-44a9-8e67-8a028d662daa", 00:07:19.002 "is_configured": true, 00:07:19.002 "data_offset": 2048, 00:07:19.003 "data_size": 63488 00:07:19.003 }, 00:07:19.003 { 00:07:19.003 "name": "BaseBdev2", 00:07:19.003 "uuid": "6ce2ae51-6b43-44b5-b798-bbde28160712", 00:07:19.003 "is_configured": true, 00:07:19.003 "data_offset": 2048, 00:07:19.003 "data_size": 63488 00:07:19.003 } 00:07:19.003 ] 00:07:19.003 } 00:07:19.003 } 00:07:19.003 }' 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:19.003 BaseBdev2' 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.003 [2024-12-07 05:34:52.286328] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:19.003 [2024-12-07 05:34:52.286399] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:19.003 [2024-12-07 05:34:52.286482] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:19.003 "name": "Existed_Raid", 00:07:19.003 "uuid": "6572ec0c-2ad3-4da6-aeb9-1f9b21ccc748", 00:07:19.003 "strip_size_kb": 64, 00:07:19.003 "state": "offline", 00:07:19.003 "raid_level": "concat", 00:07:19.003 "superblock": true, 00:07:19.003 "num_base_bdevs": 2, 00:07:19.003 "num_base_bdevs_discovered": 1, 00:07:19.003 "num_base_bdevs_operational": 1, 00:07:19.003 "base_bdevs_list": [ 00:07:19.003 { 00:07:19.003 "name": null, 00:07:19.003 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.003 "is_configured": false, 00:07:19.003 "data_offset": 0, 00:07:19.003 "data_size": 63488 00:07:19.003 }, 00:07:19.003 { 00:07:19.003 "name": "BaseBdev2", 00:07:19.003 "uuid": "6ce2ae51-6b43-44b5-b798-bbde28160712", 00:07:19.003 "is_configured": true, 00:07:19.003 "data_offset": 2048, 00:07:19.003 "data_size": 63488 00:07:19.003 } 00:07:19.003 ] 00:07:19.003 }' 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:19.003 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.573 [2024-12-07 05:34:52.732911] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:19.573 [2024-12-07 05:34:52.733012] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72984 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 72984 ']' 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 72984 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72984 00:07:19.573 killing process with pid 72984 00:07:19.573 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:19.574 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:19.574 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72984' 00:07:19.574 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 72984 00:07:19.574 [2024-12-07 05:34:52.841585] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:19.574 05:34:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 72984 00:07:19.574 [2024-12-07 05:34:52.842562] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:19.833 05:34:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:19.833 00:07:19.833 real 0m3.659s 00:07:19.833 user 0m5.769s 00:07:19.833 sys 0m0.717s 00:07:19.833 05:34:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:19.833 05:34:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.833 ************************************ 00:07:19.833 END TEST raid_state_function_test_sb 00:07:19.833 ************************************ 00:07:19.833 05:34:53 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:07:19.833 05:34:53 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:19.833 05:34:53 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:19.833 05:34:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:19.833 ************************************ 00:07:19.833 START TEST raid_superblock_test 00:07:19.833 ************************************ 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 2 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:19.833 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=73214 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 73214 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 73214 ']' 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:19.833 05:34:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.833 [2024-12-07 05:34:53.192719] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:19.833 [2024-12-07 05:34:53.192849] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73214 ] 00:07:20.093 [2024-12-07 05:34:53.348263] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:20.093 [2024-12-07 05:34:53.373499] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:20.093 [2024-12-07 05:34:53.415313] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:20.093 [2024-12-07 05:34:53.415341] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.663 05:34:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.663 malloc1 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.663 [2024-12-07 05:34:54.010582] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:20.663 [2024-12-07 05:34:54.010707] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:20.663 [2024-12-07 05:34:54.010745] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:20.663 [2024-12-07 05:34:54.010780] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:20.663 [2024-12-07 05:34:54.012928] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:20.663 [2024-12-07 05:34:54.012997] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:20.663 pt1 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.663 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.922 malloc2 00:07:20.922 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.922 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:20.922 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.922 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.922 [2024-12-07 05:34:54.043057] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:20.922 [2024-12-07 05:34:54.043168] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:20.922 [2024-12-07 05:34:54.043205] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:20.922 [2024-12-07 05:34:54.043236] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:20.922 [2024-12-07 05:34:54.045272] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:20.922 [2024-12-07 05:34:54.045342] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:20.922 pt2 00:07:20.922 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.922 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:20.922 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:20.922 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.923 [2024-12-07 05:34:54.055070] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:20.923 [2024-12-07 05:34:54.056841] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:20.923 [2024-12-07 05:34:54.056983] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:20.923 [2024-12-07 05:34:54.057003] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:20.923 [2024-12-07 05:34:54.057258] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:20.923 [2024-12-07 05:34:54.057397] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:20.923 [2024-12-07 05:34:54.057407] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:20.923 [2024-12-07 05:34:54.057525] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:20.923 "name": "raid_bdev1", 00:07:20.923 "uuid": "9a3379ad-f36c-4786-97f7-3ba17d5f9100", 00:07:20.923 "strip_size_kb": 64, 00:07:20.923 "state": "online", 00:07:20.923 "raid_level": "concat", 00:07:20.923 "superblock": true, 00:07:20.923 "num_base_bdevs": 2, 00:07:20.923 "num_base_bdevs_discovered": 2, 00:07:20.923 "num_base_bdevs_operational": 2, 00:07:20.923 "base_bdevs_list": [ 00:07:20.923 { 00:07:20.923 "name": "pt1", 00:07:20.923 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:20.923 "is_configured": true, 00:07:20.923 "data_offset": 2048, 00:07:20.923 "data_size": 63488 00:07:20.923 }, 00:07:20.923 { 00:07:20.923 "name": "pt2", 00:07:20.923 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:20.923 "is_configured": true, 00:07:20.923 "data_offset": 2048, 00:07:20.923 "data_size": 63488 00:07:20.923 } 00:07:20.923 ] 00:07:20.923 }' 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:20.923 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.183 [2024-12-07 05:34:54.454685] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:21.183 "name": "raid_bdev1", 00:07:21.183 "aliases": [ 00:07:21.183 "9a3379ad-f36c-4786-97f7-3ba17d5f9100" 00:07:21.183 ], 00:07:21.183 "product_name": "Raid Volume", 00:07:21.183 "block_size": 512, 00:07:21.183 "num_blocks": 126976, 00:07:21.183 "uuid": "9a3379ad-f36c-4786-97f7-3ba17d5f9100", 00:07:21.183 "assigned_rate_limits": { 00:07:21.183 "rw_ios_per_sec": 0, 00:07:21.183 "rw_mbytes_per_sec": 0, 00:07:21.183 "r_mbytes_per_sec": 0, 00:07:21.183 "w_mbytes_per_sec": 0 00:07:21.183 }, 00:07:21.183 "claimed": false, 00:07:21.183 "zoned": false, 00:07:21.183 "supported_io_types": { 00:07:21.183 "read": true, 00:07:21.183 "write": true, 00:07:21.183 "unmap": true, 00:07:21.183 "flush": true, 00:07:21.183 "reset": true, 00:07:21.183 "nvme_admin": false, 00:07:21.183 "nvme_io": false, 00:07:21.183 "nvme_io_md": false, 00:07:21.183 "write_zeroes": true, 00:07:21.183 "zcopy": false, 00:07:21.183 "get_zone_info": false, 00:07:21.183 "zone_management": false, 00:07:21.183 "zone_append": false, 00:07:21.183 "compare": false, 00:07:21.183 "compare_and_write": false, 00:07:21.183 "abort": false, 00:07:21.183 "seek_hole": false, 00:07:21.183 "seek_data": false, 00:07:21.183 "copy": false, 00:07:21.183 "nvme_iov_md": false 00:07:21.183 }, 00:07:21.183 "memory_domains": [ 00:07:21.183 { 00:07:21.183 "dma_device_id": "system", 00:07:21.183 "dma_device_type": 1 00:07:21.183 }, 00:07:21.183 { 00:07:21.183 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:21.183 "dma_device_type": 2 00:07:21.183 }, 00:07:21.183 { 00:07:21.183 "dma_device_id": "system", 00:07:21.183 "dma_device_type": 1 00:07:21.183 }, 00:07:21.183 { 00:07:21.183 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:21.183 "dma_device_type": 2 00:07:21.183 } 00:07:21.183 ], 00:07:21.183 "driver_specific": { 00:07:21.183 "raid": { 00:07:21.183 "uuid": "9a3379ad-f36c-4786-97f7-3ba17d5f9100", 00:07:21.183 "strip_size_kb": 64, 00:07:21.183 "state": "online", 00:07:21.183 "raid_level": "concat", 00:07:21.183 "superblock": true, 00:07:21.183 "num_base_bdevs": 2, 00:07:21.183 "num_base_bdevs_discovered": 2, 00:07:21.183 "num_base_bdevs_operational": 2, 00:07:21.183 "base_bdevs_list": [ 00:07:21.183 { 00:07:21.183 "name": "pt1", 00:07:21.183 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:21.183 "is_configured": true, 00:07:21.183 "data_offset": 2048, 00:07:21.183 "data_size": 63488 00:07:21.183 }, 00:07:21.183 { 00:07:21.183 "name": "pt2", 00:07:21.183 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:21.183 "is_configured": true, 00:07:21.183 "data_offset": 2048, 00:07:21.183 "data_size": 63488 00:07:21.183 } 00:07:21.183 ] 00:07:21.183 } 00:07:21.183 } 00:07:21.183 }' 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:21.183 pt2' 00:07:21.183 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:21.443 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.444 [2024-12-07 05:34:54.666234] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=9a3379ad-f36c-4786-97f7-3ba17d5f9100 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 9a3379ad-f36c-4786-97f7-3ba17d5f9100 ']' 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.444 [2024-12-07 05:34:54.713946] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:21.444 [2024-12-07 05:34:54.713978] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:21.444 [2024-12-07 05:34:54.714078] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:21.444 [2024-12-07 05:34:54.714143] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:21.444 [2024-12-07 05:34:54.714156] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.444 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.705 [2024-12-07 05:34:54.849749] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:21.705 [2024-12-07 05:34:54.851707] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:21.705 [2024-12-07 05:34:54.851799] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:21.705 [2024-12-07 05:34:54.851855] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:21.705 [2024-12-07 05:34:54.851875] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:21.705 [2024-12-07 05:34:54.851884] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:21.705 request: 00:07:21.705 { 00:07:21.705 "name": "raid_bdev1", 00:07:21.705 "raid_level": "concat", 00:07:21.705 "base_bdevs": [ 00:07:21.705 "malloc1", 00:07:21.705 "malloc2" 00:07:21.705 ], 00:07:21.705 "strip_size_kb": 64, 00:07:21.705 "superblock": false, 00:07:21.705 "method": "bdev_raid_create", 00:07:21.705 "req_id": 1 00:07:21.705 } 00:07:21.705 Got JSON-RPC error response 00:07:21.705 response: 00:07:21.705 { 00:07:21.705 "code": -17, 00:07:21.705 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:21.705 } 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.705 [2024-12-07 05:34:54.917580] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:21.705 [2024-12-07 05:34:54.917648] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:21.705 [2024-12-07 05:34:54.917672] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:21.705 [2024-12-07 05:34:54.917681] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:21.705 [2024-12-07 05:34:54.919824] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:21.705 [2024-12-07 05:34:54.919855] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:21.705 [2024-12-07 05:34:54.919927] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:21.705 [2024-12-07 05:34:54.919968] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:21.705 pt1 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.705 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:21.705 "name": "raid_bdev1", 00:07:21.705 "uuid": "9a3379ad-f36c-4786-97f7-3ba17d5f9100", 00:07:21.705 "strip_size_kb": 64, 00:07:21.705 "state": "configuring", 00:07:21.705 "raid_level": "concat", 00:07:21.705 "superblock": true, 00:07:21.705 "num_base_bdevs": 2, 00:07:21.705 "num_base_bdevs_discovered": 1, 00:07:21.705 "num_base_bdevs_operational": 2, 00:07:21.706 "base_bdevs_list": [ 00:07:21.706 { 00:07:21.706 "name": "pt1", 00:07:21.706 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:21.706 "is_configured": true, 00:07:21.706 "data_offset": 2048, 00:07:21.706 "data_size": 63488 00:07:21.706 }, 00:07:21.706 { 00:07:21.706 "name": null, 00:07:21.706 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:21.706 "is_configured": false, 00:07:21.706 "data_offset": 2048, 00:07:21.706 "data_size": 63488 00:07:21.706 } 00:07:21.706 ] 00:07:21.706 }' 00:07:21.706 05:34:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:21.706 05:34:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.977 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:21.977 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:21.977 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:21.977 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.978 [2024-12-07 05:34:55.324873] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:21.978 [2024-12-07 05:34:55.324982] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:21.978 [2024-12-07 05:34:55.325022] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:21.978 [2024-12-07 05:34:55.325082] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:21.978 [2024-12-07 05:34:55.325497] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:21.978 [2024-12-07 05:34:55.325552] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:21.978 [2024-12-07 05:34:55.325659] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:21.978 [2024-12-07 05:34:55.325708] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:21.978 [2024-12-07 05:34:55.325836] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:21.978 [2024-12-07 05:34:55.325875] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:21.978 [2024-12-07 05:34:55.326134] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:21.978 [2024-12-07 05:34:55.326270] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:21.978 [2024-12-07 05:34:55.326321] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:21.978 [2024-12-07 05:34:55.326457] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:21.978 pt2 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:21.978 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.236 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.236 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.236 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:22.236 "name": "raid_bdev1", 00:07:22.236 "uuid": "9a3379ad-f36c-4786-97f7-3ba17d5f9100", 00:07:22.236 "strip_size_kb": 64, 00:07:22.236 "state": "online", 00:07:22.236 "raid_level": "concat", 00:07:22.236 "superblock": true, 00:07:22.236 "num_base_bdevs": 2, 00:07:22.236 "num_base_bdevs_discovered": 2, 00:07:22.236 "num_base_bdevs_operational": 2, 00:07:22.236 "base_bdevs_list": [ 00:07:22.236 { 00:07:22.236 "name": "pt1", 00:07:22.237 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:22.237 "is_configured": true, 00:07:22.237 "data_offset": 2048, 00:07:22.237 "data_size": 63488 00:07:22.237 }, 00:07:22.237 { 00:07:22.237 "name": "pt2", 00:07:22.237 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:22.237 "is_configured": true, 00:07:22.237 "data_offset": 2048, 00:07:22.237 "data_size": 63488 00:07:22.237 } 00:07:22.237 ] 00:07:22.237 }' 00:07:22.237 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:22.237 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.498 [2024-12-07 05:34:55.812324] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:22.498 "name": "raid_bdev1", 00:07:22.498 "aliases": [ 00:07:22.498 "9a3379ad-f36c-4786-97f7-3ba17d5f9100" 00:07:22.498 ], 00:07:22.498 "product_name": "Raid Volume", 00:07:22.498 "block_size": 512, 00:07:22.498 "num_blocks": 126976, 00:07:22.498 "uuid": "9a3379ad-f36c-4786-97f7-3ba17d5f9100", 00:07:22.498 "assigned_rate_limits": { 00:07:22.498 "rw_ios_per_sec": 0, 00:07:22.498 "rw_mbytes_per_sec": 0, 00:07:22.498 "r_mbytes_per_sec": 0, 00:07:22.498 "w_mbytes_per_sec": 0 00:07:22.498 }, 00:07:22.498 "claimed": false, 00:07:22.498 "zoned": false, 00:07:22.498 "supported_io_types": { 00:07:22.498 "read": true, 00:07:22.498 "write": true, 00:07:22.498 "unmap": true, 00:07:22.498 "flush": true, 00:07:22.498 "reset": true, 00:07:22.498 "nvme_admin": false, 00:07:22.498 "nvme_io": false, 00:07:22.498 "nvme_io_md": false, 00:07:22.498 "write_zeroes": true, 00:07:22.498 "zcopy": false, 00:07:22.498 "get_zone_info": false, 00:07:22.498 "zone_management": false, 00:07:22.498 "zone_append": false, 00:07:22.498 "compare": false, 00:07:22.498 "compare_and_write": false, 00:07:22.498 "abort": false, 00:07:22.498 "seek_hole": false, 00:07:22.498 "seek_data": false, 00:07:22.498 "copy": false, 00:07:22.498 "nvme_iov_md": false 00:07:22.498 }, 00:07:22.498 "memory_domains": [ 00:07:22.498 { 00:07:22.498 "dma_device_id": "system", 00:07:22.498 "dma_device_type": 1 00:07:22.498 }, 00:07:22.498 { 00:07:22.498 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:22.498 "dma_device_type": 2 00:07:22.498 }, 00:07:22.498 { 00:07:22.498 "dma_device_id": "system", 00:07:22.498 "dma_device_type": 1 00:07:22.498 }, 00:07:22.498 { 00:07:22.498 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:22.498 "dma_device_type": 2 00:07:22.498 } 00:07:22.498 ], 00:07:22.498 "driver_specific": { 00:07:22.498 "raid": { 00:07:22.498 "uuid": "9a3379ad-f36c-4786-97f7-3ba17d5f9100", 00:07:22.498 "strip_size_kb": 64, 00:07:22.498 "state": "online", 00:07:22.498 "raid_level": "concat", 00:07:22.498 "superblock": true, 00:07:22.498 "num_base_bdevs": 2, 00:07:22.498 "num_base_bdevs_discovered": 2, 00:07:22.498 "num_base_bdevs_operational": 2, 00:07:22.498 "base_bdevs_list": [ 00:07:22.498 { 00:07:22.498 "name": "pt1", 00:07:22.498 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:22.498 "is_configured": true, 00:07:22.498 "data_offset": 2048, 00:07:22.498 "data_size": 63488 00:07:22.498 }, 00:07:22.498 { 00:07:22.498 "name": "pt2", 00:07:22.498 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:22.498 "is_configured": true, 00:07:22.498 "data_offset": 2048, 00:07:22.498 "data_size": 63488 00:07:22.498 } 00:07:22.498 ] 00:07:22.498 } 00:07:22.498 } 00:07:22.498 }' 00:07:22.498 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:22.757 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:22.757 pt2' 00:07:22.757 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:22.757 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:22.757 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:22.757 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:22.757 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:22.758 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.758 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.758 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.758 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:22.758 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:22.758 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:22.758 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:22.758 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.758 05:34:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:22.758 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.758 05:34:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.758 [2024-12-07 05:34:56.031911] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 9a3379ad-f36c-4786-97f7-3ba17d5f9100 '!=' 9a3379ad-f36c-4786-97f7-3ba17d5f9100 ']' 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 73214 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 73214 ']' 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 73214 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73214 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73214' 00:07:22.758 killing process with pid 73214 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 73214 00:07:22.758 [2024-12-07 05:34:56.103237] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:22.758 [2024-12-07 05:34:56.103324] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:22.758 [2024-12-07 05:34:56.103379] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:22.758 [2024-12-07 05:34:56.103388] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:22.758 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 73214 00:07:23.017 [2024-12-07 05:34:56.126458] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:23.017 05:34:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:23.017 00:07:23.017 real 0m3.219s 00:07:23.017 user 0m5.001s 00:07:23.017 sys 0m0.665s 00:07:23.017 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:23.017 ************************************ 00:07:23.017 END TEST raid_superblock_test 00:07:23.017 ************************************ 00:07:23.017 05:34:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.017 05:34:56 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:07:23.017 05:34:56 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:23.017 05:34:56 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:23.017 05:34:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:23.277 ************************************ 00:07:23.277 START TEST raid_read_error_test 00:07:23.277 ************************************ 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 read 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.qAR4biBq0V 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73415 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73415 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 73415 ']' 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:23.277 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:23.277 05:34:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.277 [2024-12-07 05:34:56.491028] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:23.277 [2024-12-07 05:34:56.491221] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73415 ] 00:07:23.277 [2024-12-07 05:34:56.622778] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:23.537 [2024-12-07 05:34:56.648057] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:23.537 [2024-12-07 05:34:56.690058] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:23.537 [2024-12-07 05:34:56.690091] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.107 BaseBdev1_malloc 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.107 true 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.107 [2024-12-07 05:34:57.336636] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:24.107 [2024-12-07 05:34:57.336694] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:24.107 [2024-12-07 05:34:57.336716] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:24.107 [2024-12-07 05:34:57.336725] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:24.107 [2024-12-07 05:34:57.338820] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:24.107 [2024-12-07 05:34:57.338925] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:24.107 BaseBdev1 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.107 BaseBdev2_malloc 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.107 true 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.107 [2024-12-07 05:34:57.365040] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:24.107 [2024-12-07 05:34:57.365085] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:24.107 [2024-12-07 05:34:57.365102] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:24.107 [2024-12-07 05:34:57.365117] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:24.107 [2024-12-07 05:34:57.367126] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:24.107 [2024-12-07 05:34:57.367162] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:24.107 BaseBdev2 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.107 [2024-12-07 05:34:57.373069] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:24.107 [2024-12-07 05:34:57.374932] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:24.107 [2024-12-07 05:34:57.375150] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:24.107 [2024-12-07 05:34:57.375201] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:24.107 [2024-12-07 05:34:57.375472] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:24.107 [2024-12-07 05:34:57.375667] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:24.107 [2024-12-07 05:34:57.375714] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:24.107 [2024-12-07 05:34:57.375864] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:24.107 "name": "raid_bdev1", 00:07:24.107 "uuid": "078a7e9c-d051-48c1-8618-a65fb99d04e6", 00:07:24.107 "strip_size_kb": 64, 00:07:24.107 "state": "online", 00:07:24.107 "raid_level": "concat", 00:07:24.107 "superblock": true, 00:07:24.107 "num_base_bdevs": 2, 00:07:24.107 "num_base_bdevs_discovered": 2, 00:07:24.107 "num_base_bdevs_operational": 2, 00:07:24.107 "base_bdevs_list": [ 00:07:24.107 { 00:07:24.107 "name": "BaseBdev1", 00:07:24.107 "uuid": "df8eea74-0594-57bd-b886-268d9813e800", 00:07:24.107 "is_configured": true, 00:07:24.107 "data_offset": 2048, 00:07:24.107 "data_size": 63488 00:07:24.107 }, 00:07:24.107 { 00:07:24.107 "name": "BaseBdev2", 00:07:24.107 "uuid": "e5aeb7fb-1025-518b-aae6-29128f14f00e", 00:07:24.107 "is_configured": true, 00:07:24.107 "data_offset": 2048, 00:07:24.107 "data_size": 63488 00:07:24.107 } 00:07:24.107 ] 00:07:24.107 }' 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:24.107 05:34:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.675 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:24.675 05:34:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:24.675 [2024-12-07 05:34:57.872569] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:25.613 "name": "raid_bdev1", 00:07:25.613 "uuid": "078a7e9c-d051-48c1-8618-a65fb99d04e6", 00:07:25.613 "strip_size_kb": 64, 00:07:25.613 "state": "online", 00:07:25.613 "raid_level": "concat", 00:07:25.613 "superblock": true, 00:07:25.613 "num_base_bdevs": 2, 00:07:25.613 "num_base_bdevs_discovered": 2, 00:07:25.613 "num_base_bdevs_operational": 2, 00:07:25.613 "base_bdevs_list": [ 00:07:25.613 { 00:07:25.613 "name": "BaseBdev1", 00:07:25.613 "uuid": "df8eea74-0594-57bd-b886-268d9813e800", 00:07:25.613 "is_configured": true, 00:07:25.613 "data_offset": 2048, 00:07:25.613 "data_size": 63488 00:07:25.613 }, 00:07:25.613 { 00:07:25.613 "name": "BaseBdev2", 00:07:25.613 "uuid": "e5aeb7fb-1025-518b-aae6-29128f14f00e", 00:07:25.613 "is_configured": true, 00:07:25.613 "data_offset": 2048, 00:07:25.613 "data_size": 63488 00:07:25.613 } 00:07:25.613 ] 00:07:25.613 }' 00:07:25.613 05:34:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:25.614 05:34:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.873 [2024-12-07 05:34:59.159653] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:25.873 [2024-12-07 05:34:59.159748] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:25.873 [2024-12-07 05:34:59.162256] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:25.873 [2024-12-07 05:34:59.162366] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:25.873 [2024-12-07 05:34:59.162438] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:25.873 [2024-12-07 05:34:59.162482] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.873 { 00:07:25.873 "results": [ 00:07:25.873 { 00:07:25.873 "job": "raid_bdev1", 00:07:25.873 "core_mask": "0x1", 00:07:25.873 "workload": "randrw", 00:07:25.873 "percentage": 50, 00:07:25.873 "status": "finished", 00:07:25.873 "queue_depth": 1, 00:07:25.873 "io_size": 131072, 00:07:25.873 "runtime": 1.287878, 00:07:25.873 "iops": 16958.128021443026, 00:07:25.873 "mibps": 2119.7660026803783, 00:07:25.873 "io_failed": 1, 00:07:25.873 "io_timeout": 0, 00:07:25.873 "avg_latency_us": 81.26464449597918, 00:07:25.873 "min_latency_us": 25.4882096069869, 00:07:25.873 "max_latency_us": 1380.8349344978167 00:07:25.873 } 00:07:25.873 ], 00:07:25.873 "core_count": 1 00:07:25.873 } 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73415 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 73415 ']' 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 73415 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73415 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:25.873 killing process with pid 73415 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73415' 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 73415 00:07:25.873 [2024-12-07 05:34:59.190836] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:25.873 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 73415 00:07:25.873 [2024-12-07 05:34:59.206514] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:26.134 05:34:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:26.134 05:34:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.qAR4biBq0V 00:07:26.134 05:34:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:26.134 05:34:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.78 00:07:26.134 05:34:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:07:26.134 ************************************ 00:07:26.134 END TEST raid_read_error_test 00:07:26.134 ************************************ 00:07:26.134 05:34:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:26.134 05:34:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:26.134 05:34:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.78 != \0\.\0\0 ]] 00:07:26.134 00:07:26.134 real 0m3.025s 00:07:26.134 user 0m3.824s 00:07:26.134 sys 0m0.452s 00:07:26.134 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:26.134 05:34:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.134 05:34:59 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:07:26.134 05:34:59 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:26.134 05:34:59 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:26.134 05:34:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:26.134 ************************************ 00:07:26.134 START TEST raid_write_error_test 00:07:26.134 ************************************ 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 write 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.F2n8Ajevq8 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73544 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73544 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 73544 ']' 00:07:26.134 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:26.134 05:34:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.394 [2024-12-07 05:34:59.568788] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:26.395 [2024-12-07 05:34:59.568896] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73544 ] 00:07:26.395 [2024-12-07 05:34:59.704157] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:26.395 [2024-12-07 05:34:59.730166] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:26.655 [2024-12-07 05:34:59.772368] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:26.655 [2024-12-07 05:34:59.772402] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.225 BaseBdev1_malloc 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.225 true 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.225 [2024-12-07 05:35:00.423785] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:27.225 [2024-12-07 05:35:00.423850] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:27.225 [2024-12-07 05:35:00.423892] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:27.225 [2024-12-07 05:35:00.423901] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:27.225 [2024-12-07 05:35:00.425997] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:27.225 [2024-12-07 05:35:00.426081] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:27.225 BaseBdev1 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.225 BaseBdev2_malloc 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.225 true 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.225 [2024-12-07 05:35:00.456251] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:27.225 [2024-12-07 05:35:00.456352] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:27.225 [2024-12-07 05:35:00.456374] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:27.225 [2024-12-07 05:35:00.456392] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:27.225 [2024-12-07 05:35:00.458470] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:27.225 [2024-12-07 05:35:00.458507] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:27.225 BaseBdev2 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.225 [2024-12-07 05:35:00.464296] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:27.225 [2024-12-07 05:35:00.466101] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:27.225 [2024-12-07 05:35:00.466268] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:27.225 [2024-12-07 05:35:00.466294] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:27.225 [2024-12-07 05:35:00.466554] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:27.225 [2024-12-07 05:35:00.466741] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:27.225 [2024-12-07 05:35:00.466756] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:27.225 [2024-12-07 05:35:00.466875] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.225 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.226 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.226 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:27.226 "name": "raid_bdev1", 00:07:27.226 "uuid": "365a7e33-3114-42d4-a0e9-f8ee3b266820", 00:07:27.226 "strip_size_kb": 64, 00:07:27.226 "state": "online", 00:07:27.226 "raid_level": "concat", 00:07:27.226 "superblock": true, 00:07:27.226 "num_base_bdevs": 2, 00:07:27.226 "num_base_bdevs_discovered": 2, 00:07:27.226 "num_base_bdevs_operational": 2, 00:07:27.226 "base_bdevs_list": [ 00:07:27.226 { 00:07:27.226 "name": "BaseBdev1", 00:07:27.226 "uuid": "1cf16ed1-13e2-5fb4-be73-d37f1806d1f0", 00:07:27.226 "is_configured": true, 00:07:27.226 "data_offset": 2048, 00:07:27.226 "data_size": 63488 00:07:27.226 }, 00:07:27.226 { 00:07:27.226 "name": "BaseBdev2", 00:07:27.226 "uuid": "476219ab-af2c-5bec-aa8d-cf7e403c707b", 00:07:27.226 "is_configured": true, 00:07:27.226 "data_offset": 2048, 00:07:27.226 "data_size": 63488 00:07:27.226 } 00:07:27.226 ] 00:07:27.226 }' 00:07:27.226 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:27.226 05:35:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.795 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:27.795 05:35:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:27.795 [2024-12-07 05:35:00.967814] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:28.737 "name": "raid_bdev1", 00:07:28.737 "uuid": "365a7e33-3114-42d4-a0e9-f8ee3b266820", 00:07:28.737 "strip_size_kb": 64, 00:07:28.737 "state": "online", 00:07:28.737 "raid_level": "concat", 00:07:28.737 "superblock": true, 00:07:28.737 "num_base_bdevs": 2, 00:07:28.737 "num_base_bdevs_discovered": 2, 00:07:28.737 "num_base_bdevs_operational": 2, 00:07:28.737 "base_bdevs_list": [ 00:07:28.737 { 00:07:28.737 "name": "BaseBdev1", 00:07:28.737 "uuid": "1cf16ed1-13e2-5fb4-be73-d37f1806d1f0", 00:07:28.737 "is_configured": true, 00:07:28.737 "data_offset": 2048, 00:07:28.737 "data_size": 63488 00:07:28.737 }, 00:07:28.737 { 00:07:28.737 "name": "BaseBdev2", 00:07:28.737 "uuid": "476219ab-af2c-5bec-aa8d-cf7e403c707b", 00:07:28.737 "is_configured": true, 00:07:28.737 "data_offset": 2048, 00:07:28.737 "data_size": 63488 00:07:28.737 } 00:07:28.737 ] 00:07:28.737 }' 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:28.737 05:35:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.997 05:35:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:28.997 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:28.997 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.997 [2024-12-07 05:35:02.323420] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:28.997 [2024-12-07 05:35:02.323521] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:28.997 [2024-12-07 05:35:02.326017] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:28.997 [2024-12-07 05:35:02.326076] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:28.997 [2024-12-07 05:35:02.326111] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:28.997 [2024-12-07 05:35:02.326121] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:28.997 { 00:07:28.997 "results": [ 00:07:28.997 { 00:07:28.997 "job": "raid_bdev1", 00:07:28.997 "core_mask": "0x1", 00:07:28.997 "workload": "randrw", 00:07:28.997 "percentage": 50, 00:07:28.997 "status": "finished", 00:07:28.997 "queue_depth": 1, 00:07:28.997 "io_size": 131072, 00:07:28.997 "runtime": 1.35655, 00:07:28.997 "iops": 17150.123474991706, 00:07:28.997 "mibps": 2143.765434373963, 00:07:28.997 "io_failed": 1, 00:07:28.997 "io_timeout": 0, 00:07:28.997 "avg_latency_us": 80.34115265374028, 00:07:28.997 "min_latency_us": 25.4882096069869, 00:07:28.997 "max_latency_us": 1352.216593886463 00:07:28.997 } 00:07:28.997 ], 00:07:28.997 "core_count": 1 00:07:28.997 } 00:07:28.997 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:28.997 05:35:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73544 00:07:28.997 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 73544 ']' 00:07:28.997 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 73544 00:07:28.997 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:07:28.997 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:28.997 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73544 00:07:29.257 killing process with pid 73544 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73544' 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 73544 00:07:29.257 [2024-12-07 05:35:02.368794] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 73544 00:07:29.257 [2024-12-07 05:35:02.384244] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.F2n8Ajevq8 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:07:29.257 00:07:29.257 real 0m3.123s 00:07:29.257 user 0m3.984s 00:07:29.257 sys 0m0.464s 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:29.257 ************************************ 00:07:29.257 END TEST raid_write_error_test 00:07:29.257 ************************************ 00:07:29.257 05:35:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.517 05:35:02 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:29.517 05:35:02 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:07:29.517 05:35:02 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:29.517 05:35:02 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:29.517 05:35:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:29.517 ************************************ 00:07:29.517 START TEST raid_state_function_test 00:07:29.517 ************************************ 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 false 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=73671 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73671' 00:07:29.517 Process raid pid: 73671 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 73671 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 73671 ']' 00:07:29.517 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:29.517 05:35:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.517 [2024-12-07 05:35:02.750379] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:29.517 [2024-12-07 05:35:02.750501] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:29.777 [2024-12-07 05:35:02.907847] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:29.777 [2024-12-07 05:35:02.932534] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:29.777 [2024-12-07 05:35:02.974711] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:29.777 [2024-12-07 05:35:02.974842] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.348 [2024-12-07 05:35:03.573393] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:30.348 [2024-12-07 05:35:03.573599] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:30.348 [2024-12-07 05:35:03.573634] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:30.348 [2024-12-07 05:35:03.573647] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:30.348 "name": "Existed_Raid", 00:07:30.348 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:30.348 "strip_size_kb": 0, 00:07:30.348 "state": "configuring", 00:07:30.348 "raid_level": "raid1", 00:07:30.348 "superblock": false, 00:07:30.348 "num_base_bdevs": 2, 00:07:30.348 "num_base_bdevs_discovered": 0, 00:07:30.348 "num_base_bdevs_operational": 2, 00:07:30.348 "base_bdevs_list": [ 00:07:30.348 { 00:07:30.348 "name": "BaseBdev1", 00:07:30.348 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:30.348 "is_configured": false, 00:07:30.348 "data_offset": 0, 00:07:30.348 "data_size": 0 00:07:30.348 }, 00:07:30.348 { 00:07:30.348 "name": "BaseBdev2", 00:07:30.348 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:30.348 "is_configured": false, 00:07:30.348 "data_offset": 0, 00:07:30.348 "data_size": 0 00:07:30.348 } 00:07:30.348 ] 00:07:30.348 }' 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:30.348 05:35:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.986 [2024-12-07 05:35:04.016536] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:30.986 [2024-12-07 05:35:04.016634] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.986 [2024-12-07 05:35:04.028522] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:30.986 [2024-12-07 05:35:04.028628] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:30.986 [2024-12-07 05:35:04.028660] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:30.986 [2024-12-07 05:35:04.028697] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.986 [2024-12-07 05:35:04.049412] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:30.986 BaseBdev1 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.986 [ 00:07:30.986 { 00:07:30.986 "name": "BaseBdev1", 00:07:30.986 "aliases": [ 00:07:30.986 "72f35876-4173-4e3d-973a-9f42a6beb63c" 00:07:30.986 ], 00:07:30.986 "product_name": "Malloc disk", 00:07:30.986 "block_size": 512, 00:07:30.986 "num_blocks": 65536, 00:07:30.986 "uuid": "72f35876-4173-4e3d-973a-9f42a6beb63c", 00:07:30.986 "assigned_rate_limits": { 00:07:30.986 "rw_ios_per_sec": 0, 00:07:30.986 "rw_mbytes_per_sec": 0, 00:07:30.986 "r_mbytes_per_sec": 0, 00:07:30.986 "w_mbytes_per_sec": 0 00:07:30.986 }, 00:07:30.986 "claimed": true, 00:07:30.986 "claim_type": "exclusive_write", 00:07:30.986 "zoned": false, 00:07:30.986 "supported_io_types": { 00:07:30.986 "read": true, 00:07:30.986 "write": true, 00:07:30.986 "unmap": true, 00:07:30.986 "flush": true, 00:07:30.986 "reset": true, 00:07:30.986 "nvme_admin": false, 00:07:30.986 "nvme_io": false, 00:07:30.986 "nvme_io_md": false, 00:07:30.986 "write_zeroes": true, 00:07:30.986 "zcopy": true, 00:07:30.986 "get_zone_info": false, 00:07:30.986 "zone_management": false, 00:07:30.986 "zone_append": false, 00:07:30.986 "compare": false, 00:07:30.986 "compare_and_write": false, 00:07:30.986 "abort": true, 00:07:30.986 "seek_hole": false, 00:07:30.986 "seek_data": false, 00:07:30.986 "copy": true, 00:07:30.986 "nvme_iov_md": false 00:07:30.986 }, 00:07:30.986 "memory_domains": [ 00:07:30.986 { 00:07:30.986 "dma_device_id": "system", 00:07:30.986 "dma_device_type": 1 00:07:30.986 }, 00:07:30.986 { 00:07:30.986 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:30.986 "dma_device_type": 2 00:07:30.986 } 00:07:30.986 ], 00:07:30.986 "driver_specific": {} 00:07:30.986 } 00:07:30.986 ] 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:30.986 "name": "Existed_Raid", 00:07:30.986 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:30.986 "strip_size_kb": 0, 00:07:30.986 "state": "configuring", 00:07:30.986 "raid_level": "raid1", 00:07:30.986 "superblock": false, 00:07:30.986 "num_base_bdevs": 2, 00:07:30.986 "num_base_bdevs_discovered": 1, 00:07:30.986 "num_base_bdevs_operational": 2, 00:07:30.986 "base_bdevs_list": [ 00:07:30.986 { 00:07:30.986 "name": "BaseBdev1", 00:07:30.986 "uuid": "72f35876-4173-4e3d-973a-9f42a6beb63c", 00:07:30.986 "is_configured": true, 00:07:30.986 "data_offset": 0, 00:07:30.986 "data_size": 65536 00:07:30.986 }, 00:07:30.986 { 00:07:30.986 "name": "BaseBdev2", 00:07:30.986 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:30.986 "is_configured": false, 00:07:30.986 "data_offset": 0, 00:07:30.986 "data_size": 0 00:07:30.986 } 00:07:30.986 ] 00:07:30.986 }' 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:30.986 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.247 [2024-12-07 05:35:04.480755] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:31.247 [2024-12-07 05:35:04.480808] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.247 [2024-12-07 05:35:04.488748] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:31.247 [2024-12-07 05:35:04.490591] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:31.247 [2024-12-07 05:35:04.490636] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.247 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:31.247 "name": "Existed_Raid", 00:07:31.247 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:31.247 "strip_size_kb": 0, 00:07:31.247 "state": "configuring", 00:07:31.247 "raid_level": "raid1", 00:07:31.247 "superblock": false, 00:07:31.247 "num_base_bdevs": 2, 00:07:31.247 "num_base_bdevs_discovered": 1, 00:07:31.247 "num_base_bdevs_operational": 2, 00:07:31.247 "base_bdevs_list": [ 00:07:31.247 { 00:07:31.247 "name": "BaseBdev1", 00:07:31.247 "uuid": "72f35876-4173-4e3d-973a-9f42a6beb63c", 00:07:31.247 "is_configured": true, 00:07:31.247 "data_offset": 0, 00:07:31.247 "data_size": 65536 00:07:31.247 }, 00:07:31.247 { 00:07:31.247 "name": "BaseBdev2", 00:07:31.248 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:31.248 "is_configured": false, 00:07:31.248 "data_offset": 0, 00:07:31.248 "data_size": 0 00:07:31.248 } 00:07:31.248 ] 00:07:31.248 }' 00:07:31.248 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:31.248 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.817 [2024-12-07 05:35:04.910779] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:31.817 [2024-12-07 05:35:04.910889] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:31.817 [2024-12-07 05:35:04.910916] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:07:31.817 [2024-12-07 05:35:04.911218] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:31.817 [2024-12-07 05:35:04.911405] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:31.817 [2024-12-07 05:35:04.911453] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:31.817 [2024-12-07 05:35:04.911701] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:31.817 BaseBdev2 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.817 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.817 [ 00:07:31.817 { 00:07:31.817 "name": "BaseBdev2", 00:07:31.817 "aliases": [ 00:07:31.817 "7571f960-41e4-4d36-b30a-a80ebfb0cda1" 00:07:31.817 ], 00:07:31.817 "product_name": "Malloc disk", 00:07:31.817 "block_size": 512, 00:07:31.817 "num_blocks": 65536, 00:07:31.817 "uuid": "7571f960-41e4-4d36-b30a-a80ebfb0cda1", 00:07:31.817 "assigned_rate_limits": { 00:07:31.817 "rw_ios_per_sec": 0, 00:07:31.817 "rw_mbytes_per_sec": 0, 00:07:31.817 "r_mbytes_per_sec": 0, 00:07:31.817 "w_mbytes_per_sec": 0 00:07:31.817 }, 00:07:31.818 "claimed": true, 00:07:31.818 "claim_type": "exclusive_write", 00:07:31.818 "zoned": false, 00:07:31.818 "supported_io_types": { 00:07:31.818 "read": true, 00:07:31.818 "write": true, 00:07:31.818 "unmap": true, 00:07:31.818 "flush": true, 00:07:31.818 "reset": true, 00:07:31.818 "nvme_admin": false, 00:07:31.818 "nvme_io": false, 00:07:31.818 "nvme_io_md": false, 00:07:31.818 "write_zeroes": true, 00:07:31.818 "zcopy": true, 00:07:31.818 "get_zone_info": false, 00:07:31.818 "zone_management": false, 00:07:31.818 "zone_append": false, 00:07:31.818 "compare": false, 00:07:31.818 "compare_and_write": false, 00:07:31.818 "abort": true, 00:07:31.818 "seek_hole": false, 00:07:31.818 "seek_data": false, 00:07:31.818 "copy": true, 00:07:31.818 "nvme_iov_md": false 00:07:31.818 }, 00:07:31.818 "memory_domains": [ 00:07:31.818 { 00:07:31.818 "dma_device_id": "system", 00:07:31.818 "dma_device_type": 1 00:07:31.818 }, 00:07:31.818 { 00:07:31.818 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:31.818 "dma_device_type": 2 00:07:31.818 } 00:07:31.818 ], 00:07:31.818 "driver_specific": {} 00:07:31.818 } 00:07:31.818 ] 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:31.818 "name": "Existed_Raid", 00:07:31.818 "uuid": "8a08b45e-6eee-400c-bdbf-db74da3a99a6", 00:07:31.818 "strip_size_kb": 0, 00:07:31.818 "state": "online", 00:07:31.818 "raid_level": "raid1", 00:07:31.818 "superblock": false, 00:07:31.818 "num_base_bdevs": 2, 00:07:31.818 "num_base_bdevs_discovered": 2, 00:07:31.818 "num_base_bdevs_operational": 2, 00:07:31.818 "base_bdevs_list": [ 00:07:31.818 { 00:07:31.818 "name": "BaseBdev1", 00:07:31.818 "uuid": "72f35876-4173-4e3d-973a-9f42a6beb63c", 00:07:31.818 "is_configured": true, 00:07:31.818 "data_offset": 0, 00:07:31.818 "data_size": 65536 00:07:31.818 }, 00:07:31.818 { 00:07:31.818 "name": "BaseBdev2", 00:07:31.818 "uuid": "7571f960-41e4-4d36-b30a-a80ebfb0cda1", 00:07:31.818 "is_configured": true, 00:07:31.818 "data_offset": 0, 00:07:31.818 "data_size": 65536 00:07:31.818 } 00:07:31.818 ] 00:07:31.818 }' 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:31.818 05:35:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.077 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:32.077 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:32.077 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:32.077 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:32.077 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:32.077 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:32.077 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:32.077 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:32.077 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.077 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.077 [2024-12-07 05:35:05.406393] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:32.077 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:32.334 "name": "Existed_Raid", 00:07:32.334 "aliases": [ 00:07:32.334 "8a08b45e-6eee-400c-bdbf-db74da3a99a6" 00:07:32.334 ], 00:07:32.334 "product_name": "Raid Volume", 00:07:32.334 "block_size": 512, 00:07:32.334 "num_blocks": 65536, 00:07:32.334 "uuid": "8a08b45e-6eee-400c-bdbf-db74da3a99a6", 00:07:32.334 "assigned_rate_limits": { 00:07:32.334 "rw_ios_per_sec": 0, 00:07:32.334 "rw_mbytes_per_sec": 0, 00:07:32.334 "r_mbytes_per_sec": 0, 00:07:32.334 "w_mbytes_per_sec": 0 00:07:32.334 }, 00:07:32.334 "claimed": false, 00:07:32.334 "zoned": false, 00:07:32.334 "supported_io_types": { 00:07:32.334 "read": true, 00:07:32.334 "write": true, 00:07:32.334 "unmap": false, 00:07:32.334 "flush": false, 00:07:32.334 "reset": true, 00:07:32.334 "nvme_admin": false, 00:07:32.334 "nvme_io": false, 00:07:32.334 "nvme_io_md": false, 00:07:32.334 "write_zeroes": true, 00:07:32.334 "zcopy": false, 00:07:32.334 "get_zone_info": false, 00:07:32.334 "zone_management": false, 00:07:32.334 "zone_append": false, 00:07:32.334 "compare": false, 00:07:32.334 "compare_and_write": false, 00:07:32.334 "abort": false, 00:07:32.334 "seek_hole": false, 00:07:32.334 "seek_data": false, 00:07:32.334 "copy": false, 00:07:32.334 "nvme_iov_md": false 00:07:32.334 }, 00:07:32.334 "memory_domains": [ 00:07:32.334 { 00:07:32.334 "dma_device_id": "system", 00:07:32.334 "dma_device_type": 1 00:07:32.334 }, 00:07:32.334 { 00:07:32.334 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:32.334 "dma_device_type": 2 00:07:32.334 }, 00:07:32.334 { 00:07:32.334 "dma_device_id": "system", 00:07:32.334 "dma_device_type": 1 00:07:32.334 }, 00:07:32.334 { 00:07:32.334 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:32.334 "dma_device_type": 2 00:07:32.334 } 00:07:32.334 ], 00:07:32.334 "driver_specific": { 00:07:32.334 "raid": { 00:07:32.334 "uuid": "8a08b45e-6eee-400c-bdbf-db74da3a99a6", 00:07:32.334 "strip_size_kb": 0, 00:07:32.334 "state": "online", 00:07:32.334 "raid_level": "raid1", 00:07:32.334 "superblock": false, 00:07:32.334 "num_base_bdevs": 2, 00:07:32.334 "num_base_bdevs_discovered": 2, 00:07:32.334 "num_base_bdevs_operational": 2, 00:07:32.334 "base_bdevs_list": [ 00:07:32.334 { 00:07:32.334 "name": "BaseBdev1", 00:07:32.334 "uuid": "72f35876-4173-4e3d-973a-9f42a6beb63c", 00:07:32.334 "is_configured": true, 00:07:32.334 "data_offset": 0, 00:07:32.334 "data_size": 65536 00:07:32.334 }, 00:07:32.334 { 00:07:32.334 "name": "BaseBdev2", 00:07:32.334 "uuid": "7571f960-41e4-4d36-b30a-a80ebfb0cda1", 00:07:32.334 "is_configured": true, 00:07:32.334 "data_offset": 0, 00:07:32.334 "data_size": 65536 00:07:32.334 } 00:07:32.334 ] 00:07:32.334 } 00:07:32.334 } 00:07:32.334 }' 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:32.334 BaseBdev2' 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.334 [2024-12-07 05:35:05.641768] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:32.334 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:32.335 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.335 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.335 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.592 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:32.592 "name": "Existed_Raid", 00:07:32.592 "uuid": "8a08b45e-6eee-400c-bdbf-db74da3a99a6", 00:07:32.592 "strip_size_kb": 0, 00:07:32.592 "state": "online", 00:07:32.592 "raid_level": "raid1", 00:07:32.592 "superblock": false, 00:07:32.592 "num_base_bdevs": 2, 00:07:32.592 "num_base_bdevs_discovered": 1, 00:07:32.592 "num_base_bdevs_operational": 1, 00:07:32.592 "base_bdevs_list": [ 00:07:32.592 { 00:07:32.592 "name": null, 00:07:32.592 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:32.592 "is_configured": false, 00:07:32.592 "data_offset": 0, 00:07:32.592 "data_size": 65536 00:07:32.592 }, 00:07:32.592 { 00:07:32.592 "name": "BaseBdev2", 00:07:32.592 "uuid": "7571f960-41e4-4d36-b30a-a80ebfb0cda1", 00:07:32.592 "is_configured": true, 00:07:32.592 "data_offset": 0, 00:07:32.592 "data_size": 65536 00:07:32.592 } 00:07:32.592 ] 00:07:32.592 }' 00:07:32.592 05:35:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:32.592 05:35:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.850 [2024-12-07 05:35:06.096072] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:32.850 [2024-12-07 05:35:06.096227] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:32.850 [2024-12-07 05:35:06.107782] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:32.850 [2024-12-07 05:35:06.107829] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:32.850 [2024-12-07 05:35:06.107841] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 73671 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 73671 ']' 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 73671 00:07:32.850 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:32.851 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:32.851 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73671 00:07:32.851 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:32.851 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:32.851 killing process with pid 73671 00:07:32.851 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73671' 00:07:32.851 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 73671 00:07:32.851 [2024-12-07 05:35:06.191995] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:32.851 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 73671 00:07:32.851 [2024-12-07 05:35:06.192980] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:33.109 ************************************ 00:07:33.109 END TEST raid_state_function_test 00:07:33.109 ************************************ 00:07:33.109 05:35:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:33.109 00:07:33.109 real 0m3.742s 00:07:33.109 user 0m5.934s 00:07:33.109 sys 0m0.737s 00:07:33.109 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:33.109 05:35:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.109 05:35:06 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:07:33.109 05:35:06 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:33.109 05:35:06 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:33.109 05:35:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:33.109 ************************************ 00:07:33.109 START TEST raid_state_function_test_sb 00:07:33.109 ************************************ 00:07:33.109 05:35:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:07:33.109 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:07:33.109 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:33.109 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:33.109 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=73907 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73907' 00:07:33.368 Process raid pid: 73907 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 73907 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 73907 ']' 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:33.368 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:33.368 05:35:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.368 [2024-12-07 05:35:06.560023] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:33.368 [2024-12-07 05:35:06.560135] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:33.368 [2024-12-07 05:35:06.714000] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:33.627 [2024-12-07 05:35:06.738524] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:33.627 [2024-12-07 05:35:06.779772] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:33.627 [2024-12-07 05:35:06.779901] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.197 [2024-12-07 05:35:07.382050] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:34.197 [2024-12-07 05:35:07.382110] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:34.197 [2024-12-07 05:35:07.382128] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:34.197 [2024-12-07 05:35:07.382138] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:34.197 "name": "Existed_Raid", 00:07:34.197 "uuid": "2a03cc19-5900-4f26-bab8-6f1feccb65bc", 00:07:34.197 "strip_size_kb": 0, 00:07:34.197 "state": "configuring", 00:07:34.197 "raid_level": "raid1", 00:07:34.197 "superblock": true, 00:07:34.197 "num_base_bdevs": 2, 00:07:34.197 "num_base_bdevs_discovered": 0, 00:07:34.197 "num_base_bdevs_operational": 2, 00:07:34.197 "base_bdevs_list": [ 00:07:34.197 { 00:07:34.197 "name": "BaseBdev1", 00:07:34.197 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.197 "is_configured": false, 00:07:34.197 "data_offset": 0, 00:07:34.197 "data_size": 0 00:07:34.197 }, 00:07:34.197 { 00:07:34.197 "name": "BaseBdev2", 00:07:34.197 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.197 "is_configured": false, 00:07:34.197 "data_offset": 0, 00:07:34.197 "data_size": 0 00:07:34.197 } 00:07:34.197 ] 00:07:34.197 }' 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:34.197 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.768 [2024-12-07 05:35:07.833206] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:34.768 [2024-12-07 05:35:07.833248] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.768 [2024-12-07 05:35:07.841210] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:34.768 [2024-12-07 05:35:07.841257] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:34.768 [2024-12-07 05:35:07.841267] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:34.768 [2024-12-07 05:35:07.841290] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.768 [2024-12-07 05:35:07.857824] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:34.768 BaseBdev1 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.768 [ 00:07:34.768 { 00:07:34.768 "name": "BaseBdev1", 00:07:34.768 "aliases": [ 00:07:34.768 "9d57f856-41af-4dad-819f-825e5f05e77e" 00:07:34.768 ], 00:07:34.768 "product_name": "Malloc disk", 00:07:34.768 "block_size": 512, 00:07:34.768 "num_blocks": 65536, 00:07:34.768 "uuid": "9d57f856-41af-4dad-819f-825e5f05e77e", 00:07:34.768 "assigned_rate_limits": { 00:07:34.768 "rw_ios_per_sec": 0, 00:07:34.768 "rw_mbytes_per_sec": 0, 00:07:34.768 "r_mbytes_per_sec": 0, 00:07:34.768 "w_mbytes_per_sec": 0 00:07:34.768 }, 00:07:34.768 "claimed": true, 00:07:34.768 "claim_type": "exclusive_write", 00:07:34.768 "zoned": false, 00:07:34.768 "supported_io_types": { 00:07:34.768 "read": true, 00:07:34.768 "write": true, 00:07:34.768 "unmap": true, 00:07:34.768 "flush": true, 00:07:34.768 "reset": true, 00:07:34.768 "nvme_admin": false, 00:07:34.768 "nvme_io": false, 00:07:34.768 "nvme_io_md": false, 00:07:34.768 "write_zeroes": true, 00:07:34.768 "zcopy": true, 00:07:34.768 "get_zone_info": false, 00:07:34.768 "zone_management": false, 00:07:34.768 "zone_append": false, 00:07:34.768 "compare": false, 00:07:34.768 "compare_and_write": false, 00:07:34.768 "abort": true, 00:07:34.768 "seek_hole": false, 00:07:34.768 "seek_data": false, 00:07:34.768 "copy": true, 00:07:34.768 "nvme_iov_md": false 00:07:34.768 }, 00:07:34.768 "memory_domains": [ 00:07:34.768 { 00:07:34.768 "dma_device_id": "system", 00:07:34.768 "dma_device_type": 1 00:07:34.768 }, 00:07:34.768 { 00:07:34.768 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:34.768 "dma_device_type": 2 00:07:34.768 } 00:07:34.768 ], 00:07:34.768 "driver_specific": {} 00:07:34.768 } 00:07:34.768 ] 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:34.768 "name": "Existed_Raid", 00:07:34.768 "uuid": "e8d3c299-e6ea-482d-af50-469b04a7e0d3", 00:07:34.768 "strip_size_kb": 0, 00:07:34.768 "state": "configuring", 00:07:34.768 "raid_level": "raid1", 00:07:34.768 "superblock": true, 00:07:34.768 "num_base_bdevs": 2, 00:07:34.768 "num_base_bdevs_discovered": 1, 00:07:34.768 "num_base_bdevs_operational": 2, 00:07:34.768 "base_bdevs_list": [ 00:07:34.768 { 00:07:34.768 "name": "BaseBdev1", 00:07:34.768 "uuid": "9d57f856-41af-4dad-819f-825e5f05e77e", 00:07:34.768 "is_configured": true, 00:07:34.768 "data_offset": 2048, 00:07:34.768 "data_size": 63488 00:07:34.768 }, 00:07:34.768 { 00:07:34.768 "name": "BaseBdev2", 00:07:34.768 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.768 "is_configured": false, 00:07:34.768 "data_offset": 0, 00:07:34.768 "data_size": 0 00:07:34.768 } 00:07:34.768 ] 00:07:34.768 }' 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:34.768 05:35:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.029 [2024-12-07 05:35:08.345036] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:35.029 [2024-12-07 05:35:08.345125] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.029 [2024-12-07 05:35:08.353057] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:35.029 [2024-12-07 05:35:08.354982] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:35.029 [2024-12-07 05:35:08.355057] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:35.029 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:35.030 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.030 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.030 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:35.030 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.030 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.290 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:35.290 "name": "Existed_Raid", 00:07:35.290 "uuid": "16ecfd07-2c38-4e66-9012-3be38f13ed72", 00:07:35.290 "strip_size_kb": 0, 00:07:35.290 "state": "configuring", 00:07:35.290 "raid_level": "raid1", 00:07:35.290 "superblock": true, 00:07:35.290 "num_base_bdevs": 2, 00:07:35.290 "num_base_bdevs_discovered": 1, 00:07:35.290 "num_base_bdevs_operational": 2, 00:07:35.290 "base_bdevs_list": [ 00:07:35.290 { 00:07:35.290 "name": "BaseBdev1", 00:07:35.290 "uuid": "9d57f856-41af-4dad-819f-825e5f05e77e", 00:07:35.290 "is_configured": true, 00:07:35.290 "data_offset": 2048, 00:07:35.290 "data_size": 63488 00:07:35.290 }, 00:07:35.290 { 00:07:35.290 "name": "BaseBdev2", 00:07:35.290 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.290 "is_configured": false, 00:07:35.290 "data_offset": 0, 00:07:35.290 "data_size": 0 00:07:35.290 } 00:07:35.290 ] 00:07:35.290 }' 00:07:35.290 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:35.290 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.550 [2024-12-07 05:35:08.783204] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:35.550 [2024-12-07 05:35:08.783511] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:35.550 [2024-12-07 05:35:08.783567] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:35.550 BaseBdev2 00:07:35.550 [2024-12-07 05:35:08.783899] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:35.550 [2024-12-07 05:35:08.784099] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:35.550 [2024-12-07 05:35:08.784146] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:35.550 [2024-12-07 05:35:08.784307] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.550 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.550 [ 00:07:35.550 { 00:07:35.550 "name": "BaseBdev2", 00:07:35.550 "aliases": [ 00:07:35.550 "5aabbc32-0ef3-4b17-a1cd-bd5e8da1339f" 00:07:35.550 ], 00:07:35.550 "product_name": "Malloc disk", 00:07:35.550 "block_size": 512, 00:07:35.550 "num_blocks": 65536, 00:07:35.550 "uuid": "5aabbc32-0ef3-4b17-a1cd-bd5e8da1339f", 00:07:35.550 "assigned_rate_limits": { 00:07:35.550 "rw_ios_per_sec": 0, 00:07:35.550 "rw_mbytes_per_sec": 0, 00:07:35.550 "r_mbytes_per_sec": 0, 00:07:35.550 "w_mbytes_per_sec": 0 00:07:35.550 }, 00:07:35.550 "claimed": true, 00:07:35.550 "claim_type": "exclusive_write", 00:07:35.550 "zoned": false, 00:07:35.550 "supported_io_types": { 00:07:35.550 "read": true, 00:07:35.550 "write": true, 00:07:35.550 "unmap": true, 00:07:35.550 "flush": true, 00:07:35.550 "reset": true, 00:07:35.550 "nvme_admin": false, 00:07:35.550 "nvme_io": false, 00:07:35.550 "nvme_io_md": false, 00:07:35.551 "write_zeroes": true, 00:07:35.551 "zcopy": true, 00:07:35.551 "get_zone_info": false, 00:07:35.551 "zone_management": false, 00:07:35.551 "zone_append": false, 00:07:35.551 "compare": false, 00:07:35.551 "compare_and_write": false, 00:07:35.551 "abort": true, 00:07:35.551 "seek_hole": false, 00:07:35.551 "seek_data": false, 00:07:35.551 "copy": true, 00:07:35.551 "nvme_iov_md": false 00:07:35.551 }, 00:07:35.551 "memory_domains": [ 00:07:35.551 { 00:07:35.551 "dma_device_id": "system", 00:07:35.551 "dma_device_type": 1 00:07:35.551 }, 00:07:35.551 { 00:07:35.551 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.551 "dma_device_type": 2 00:07:35.551 } 00:07:35.551 ], 00:07:35.551 "driver_specific": {} 00:07:35.551 } 00:07:35.551 ] 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:35.551 "name": "Existed_Raid", 00:07:35.551 "uuid": "16ecfd07-2c38-4e66-9012-3be38f13ed72", 00:07:35.551 "strip_size_kb": 0, 00:07:35.551 "state": "online", 00:07:35.551 "raid_level": "raid1", 00:07:35.551 "superblock": true, 00:07:35.551 "num_base_bdevs": 2, 00:07:35.551 "num_base_bdevs_discovered": 2, 00:07:35.551 "num_base_bdevs_operational": 2, 00:07:35.551 "base_bdevs_list": [ 00:07:35.551 { 00:07:35.551 "name": "BaseBdev1", 00:07:35.551 "uuid": "9d57f856-41af-4dad-819f-825e5f05e77e", 00:07:35.551 "is_configured": true, 00:07:35.551 "data_offset": 2048, 00:07:35.551 "data_size": 63488 00:07:35.551 }, 00:07:35.551 { 00:07:35.551 "name": "BaseBdev2", 00:07:35.551 "uuid": "5aabbc32-0ef3-4b17-a1cd-bd5e8da1339f", 00:07:35.551 "is_configured": true, 00:07:35.551 "data_offset": 2048, 00:07:35.551 "data_size": 63488 00:07:35.551 } 00:07:35.551 ] 00:07:35.551 }' 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:35.551 05:35:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.121 [2024-12-07 05:35:09.222778] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.121 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:36.121 "name": "Existed_Raid", 00:07:36.121 "aliases": [ 00:07:36.121 "16ecfd07-2c38-4e66-9012-3be38f13ed72" 00:07:36.121 ], 00:07:36.121 "product_name": "Raid Volume", 00:07:36.121 "block_size": 512, 00:07:36.121 "num_blocks": 63488, 00:07:36.121 "uuid": "16ecfd07-2c38-4e66-9012-3be38f13ed72", 00:07:36.121 "assigned_rate_limits": { 00:07:36.121 "rw_ios_per_sec": 0, 00:07:36.121 "rw_mbytes_per_sec": 0, 00:07:36.121 "r_mbytes_per_sec": 0, 00:07:36.121 "w_mbytes_per_sec": 0 00:07:36.121 }, 00:07:36.121 "claimed": false, 00:07:36.121 "zoned": false, 00:07:36.121 "supported_io_types": { 00:07:36.121 "read": true, 00:07:36.121 "write": true, 00:07:36.121 "unmap": false, 00:07:36.121 "flush": false, 00:07:36.121 "reset": true, 00:07:36.121 "nvme_admin": false, 00:07:36.121 "nvme_io": false, 00:07:36.121 "nvme_io_md": false, 00:07:36.121 "write_zeroes": true, 00:07:36.121 "zcopy": false, 00:07:36.121 "get_zone_info": false, 00:07:36.121 "zone_management": false, 00:07:36.121 "zone_append": false, 00:07:36.121 "compare": false, 00:07:36.121 "compare_and_write": false, 00:07:36.121 "abort": false, 00:07:36.121 "seek_hole": false, 00:07:36.121 "seek_data": false, 00:07:36.121 "copy": false, 00:07:36.121 "nvme_iov_md": false 00:07:36.121 }, 00:07:36.121 "memory_domains": [ 00:07:36.121 { 00:07:36.121 "dma_device_id": "system", 00:07:36.121 "dma_device_type": 1 00:07:36.121 }, 00:07:36.121 { 00:07:36.121 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:36.121 "dma_device_type": 2 00:07:36.122 }, 00:07:36.122 { 00:07:36.122 "dma_device_id": "system", 00:07:36.122 "dma_device_type": 1 00:07:36.122 }, 00:07:36.122 { 00:07:36.122 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:36.122 "dma_device_type": 2 00:07:36.122 } 00:07:36.122 ], 00:07:36.122 "driver_specific": { 00:07:36.122 "raid": { 00:07:36.122 "uuid": "16ecfd07-2c38-4e66-9012-3be38f13ed72", 00:07:36.122 "strip_size_kb": 0, 00:07:36.122 "state": "online", 00:07:36.122 "raid_level": "raid1", 00:07:36.122 "superblock": true, 00:07:36.122 "num_base_bdevs": 2, 00:07:36.122 "num_base_bdevs_discovered": 2, 00:07:36.122 "num_base_bdevs_operational": 2, 00:07:36.122 "base_bdevs_list": [ 00:07:36.122 { 00:07:36.122 "name": "BaseBdev1", 00:07:36.122 "uuid": "9d57f856-41af-4dad-819f-825e5f05e77e", 00:07:36.122 "is_configured": true, 00:07:36.122 "data_offset": 2048, 00:07:36.122 "data_size": 63488 00:07:36.122 }, 00:07:36.122 { 00:07:36.122 "name": "BaseBdev2", 00:07:36.122 "uuid": "5aabbc32-0ef3-4b17-a1cd-bd5e8da1339f", 00:07:36.122 "is_configured": true, 00:07:36.122 "data_offset": 2048, 00:07:36.122 "data_size": 63488 00:07:36.122 } 00:07:36.122 ] 00:07:36.122 } 00:07:36.122 } 00:07:36.122 }' 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:36.122 BaseBdev2' 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.122 [2024-12-07 05:35:09.454515] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:36.122 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.382 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:36.382 "name": "Existed_Raid", 00:07:36.382 "uuid": "16ecfd07-2c38-4e66-9012-3be38f13ed72", 00:07:36.382 "strip_size_kb": 0, 00:07:36.382 "state": "online", 00:07:36.382 "raid_level": "raid1", 00:07:36.382 "superblock": true, 00:07:36.382 "num_base_bdevs": 2, 00:07:36.382 "num_base_bdevs_discovered": 1, 00:07:36.382 "num_base_bdevs_operational": 1, 00:07:36.382 "base_bdevs_list": [ 00:07:36.382 { 00:07:36.382 "name": null, 00:07:36.382 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.382 "is_configured": false, 00:07:36.382 "data_offset": 0, 00:07:36.382 "data_size": 63488 00:07:36.382 }, 00:07:36.382 { 00:07:36.382 "name": "BaseBdev2", 00:07:36.382 "uuid": "5aabbc32-0ef3-4b17-a1cd-bd5e8da1339f", 00:07:36.382 "is_configured": true, 00:07:36.382 "data_offset": 2048, 00:07:36.382 "data_size": 63488 00:07:36.382 } 00:07:36.382 ] 00:07:36.382 }' 00:07:36.382 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:36.382 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.642 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:36.642 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.643 [2024-12-07 05:35:09.960964] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:36.643 [2024-12-07 05:35:09.961114] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:36.643 [2024-12-07 05:35:09.972649] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:36.643 [2024-12-07 05:35:09.972769] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:36.643 [2024-12-07 05:35:09.972812] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:36.643 05:35:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 73907 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 73907 ']' 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 73907 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73907 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73907' 00:07:36.904 killing process with pid 73907 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 73907 00:07:36.904 [2024-12-07 05:35:10.074853] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:36.904 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 73907 00:07:36.904 [2024-12-07 05:35:10.075829] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:37.164 05:35:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:37.164 00:07:37.164 real 0m3.813s 00:07:37.164 user 0m6.096s 00:07:37.164 sys 0m0.700s 00:07:37.164 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:37.164 ************************************ 00:07:37.164 END TEST raid_state_function_test_sb 00:07:37.164 ************************************ 00:07:37.164 05:35:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:37.164 05:35:10 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:07:37.164 05:35:10 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:37.164 05:35:10 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:37.164 05:35:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:37.164 ************************************ 00:07:37.164 START TEST raid_superblock_test 00:07:37.164 ************************************ 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:07:37.164 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=74144 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 74144 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 74144 ']' 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:37.164 05:35:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.164 [2024-12-07 05:35:10.435556] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:37.164 [2024-12-07 05:35:10.436133] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74144 ] 00:07:37.424 [2024-12-07 05:35:10.567741] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:37.424 [2024-12-07 05:35:10.591974] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:37.424 [2024-12-07 05:35:10.634205] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:37.424 [2024-12-07 05:35:10.634236] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.994 malloc1 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.994 [2024-12-07 05:35:11.273102] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:37.994 [2024-12-07 05:35:11.273208] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:37.994 [2024-12-07 05:35:11.273271] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:37.994 [2024-12-07 05:35:11.273305] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:37.994 [2024-12-07 05:35:11.275363] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:37.994 [2024-12-07 05:35:11.275439] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:37.994 pt1 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.994 malloc2 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.994 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.994 [2024-12-07 05:35:11.301550] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:37.995 [2024-12-07 05:35:11.301608] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:37.995 [2024-12-07 05:35:11.301643] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:37.995 [2024-12-07 05:35:11.301655] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:37.995 [2024-12-07 05:35:11.303726] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:37.995 [2024-12-07 05:35:11.303763] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:37.995 pt2 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.995 [2024-12-07 05:35:11.313569] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:37.995 [2024-12-07 05:35:11.315405] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:37.995 [2024-12-07 05:35:11.315543] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:37.995 [2024-12-07 05:35:11.315558] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:37.995 [2024-12-07 05:35:11.315825] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:37.995 [2024-12-07 05:35:11.315965] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:37.995 [2024-12-07 05:35:11.315982] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:37.995 [2024-12-07 05:35:11.316104] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.995 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.254 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:38.254 "name": "raid_bdev1", 00:07:38.254 "uuid": "ad9f0507-a8fe-46a8-9da8-600ef57a8e24", 00:07:38.254 "strip_size_kb": 0, 00:07:38.254 "state": "online", 00:07:38.254 "raid_level": "raid1", 00:07:38.254 "superblock": true, 00:07:38.254 "num_base_bdevs": 2, 00:07:38.254 "num_base_bdevs_discovered": 2, 00:07:38.254 "num_base_bdevs_operational": 2, 00:07:38.254 "base_bdevs_list": [ 00:07:38.254 { 00:07:38.254 "name": "pt1", 00:07:38.254 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:38.254 "is_configured": true, 00:07:38.254 "data_offset": 2048, 00:07:38.254 "data_size": 63488 00:07:38.254 }, 00:07:38.254 { 00:07:38.254 "name": "pt2", 00:07:38.254 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:38.254 "is_configured": true, 00:07:38.254 "data_offset": 2048, 00:07:38.254 "data_size": 63488 00:07:38.254 } 00:07:38.254 ] 00:07:38.254 }' 00:07:38.254 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:38.254 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.513 [2024-12-07 05:35:11.745105] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:38.513 "name": "raid_bdev1", 00:07:38.513 "aliases": [ 00:07:38.513 "ad9f0507-a8fe-46a8-9da8-600ef57a8e24" 00:07:38.513 ], 00:07:38.513 "product_name": "Raid Volume", 00:07:38.513 "block_size": 512, 00:07:38.513 "num_blocks": 63488, 00:07:38.513 "uuid": "ad9f0507-a8fe-46a8-9da8-600ef57a8e24", 00:07:38.513 "assigned_rate_limits": { 00:07:38.513 "rw_ios_per_sec": 0, 00:07:38.513 "rw_mbytes_per_sec": 0, 00:07:38.513 "r_mbytes_per_sec": 0, 00:07:38.513 "w_mbytes_per_sec": 0 00:07:38.513 }, 00:07:38.513 "claimed": false, 00:07:38.513 "zoned": false, 00:07:38.513 "supported_io_types": { 00:07:38.513 "read": true, 00:07:38.513 "write": true, 00:07:38.513 "unmap": false, 00:07:38.513 "flush": false, 00:07:38.513 "reset": true, 00:07:38.513 "nvme_admin": false, 00:07:38.513 "nvme_io": false, 00:07:38.513 "nvme_io_md": false, 00:07:38.513 "write_zeroes": true, 00:07:38.513 "zcopy": false, 00:07:38.513 "get_zone_info": false, 00:07:38.513 "zone_management": false, 00:07:38.513 "zone_append": false, 00:07:38.513 "compare": false, 00:07:38.513 "compare_and_write": false, 00:07:38.513 "abort": false, 00:07:38.513 "seek_hole": false, 00:07:38.513 "seek_data": false, 00:07:38.513 "copy": false, 00:07:38.513 "nvme_iov_md": false 00:07:38.513 }, 00:07:38.513 "memory_domains": [ 00:07:38.513 { 00:07:38.513 "dma_device_id": "system", 00:07:38.513 "dma_device_type": 1 00:07:38.513 }, 00:07:38.513 { 00:07:38.513 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:38.513 "dma_device_type": 2 00:07:38.513 }, 00:07:38.513 { 00:07:38.513 "dma_device_id": "system", 00:07:38.513 "dma_device_type": 1 00:07:38.513 }, 00:07:38.513 { 00:07:38.513 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:38.513 "dma_device_type": 2 00:07:38.513 } 00:07:38.513 ], 00:07:38.513 "driver_specific": { 00:07:38.513 "raid": { 00:07:38.513 "uuid": "ad9f0507-a8fe-46a8-9da8-600ef57a8e24", 00:07:38.513 "strip_size_kb": 0, 00:07:38.513 "state": "online", 00:07:38.513 "raid_level": "raid1", 00:07:38.513 "superblock": true, 00:07:38.513 "num_base_bdevs": 2, 00:07:38.513 "num_base_bdevs_discovered": 2, 00:07:38.513 "num_base_bdevs_operational": 2, 00:07:38.513 "base_bdevs_list": [ 00:07:38.513 { 00:07:38.513 "name": "pt1", 00:07:38.513 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:38.513 "is_configured": true, 00:07:38.513 "data_offset": 2048, 00:07:38.513 "data_size": 63488 00:07:38.513 }, 00:07:38.513 { 00:07:38.513 "name": "pt2", 00:07:38.513 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:38.513 "is_configured": true, 00:07:38.513 "data_offset": 2048, 00:07:38.513 "data_size": 63488 00:07:38.513 } 00:07:38.513 ] 00:07:38.513 } 00:07:38.513 } 00:07:38.513 }' 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:38.513 pt2' 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:38.513 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.773 [2024-12-07 05:35:11.964610] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:38.773 05:35:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=ad9f0507-a8fe-46a8-9da8-600ef57a8e24 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z ad9f0507-a8fe-46a8-9da8-600ef57a8e24 ']' 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.773 [2024-12-07 05:35:12.012322] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:38.773 [2024-12-07 05:35:12.012391] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:38.773 [2024-12-07 05:35:12.012475] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:38.773 [2024-12-07 05:35:12.012560] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:38.773 [2024-12-07 05:35:12.012575] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.773 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.773 [2024-12-07 05:35:12.128139] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:38.774 [2024-12-07 05:35:12.130045] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:38.774 [2024-12-07 05:35:12.130172] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:38.774 [2024-12-07 05:35:12.130277] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:38.774 [2024-12-07 05:35:12.130339] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:38.774 [2024-12-07 05:35:12.130350] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:38.774 request: 00:07:38.774 { 00:07:38.774 "name": "raid_bdev1", 00:07:38.774 "raid_level": "raid1", 00:07:38.774 "base_bdevs": [ 00:07:38.774 "malloc1", 00:07:38.774 "malloc2" 00:07:38.774 ], 00:07:38.774 "superblock": false, 00:07:38.774 "method": "bdev_raid_create", 00:07:38.774 "req_id": 1 00:07:38.774 } 00:07:38.774 Got JSON-RPC error response 00:07:38.774 response: 00:07:38.774 { 00:07:38.774 "code": -17, 00:07:38.774 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:38.774 } 00:07:38.774 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:07:38.774 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:07:38.774 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:38.774 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:38.774 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:39.040 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:39.040 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:39.040 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.040 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.040 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.040 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:39.040 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:39.040 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:39.040 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.040 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.040 [2024-12-07 05:35:12.192010] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:39.040 [2024-12-07 05:35:12.192103] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:39.041 [2024-12-07 05:35:12.192141] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:39.041 [2024-12-07 05:35:12.192168] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:39.041 [2024-12-07 05:35:12.194277] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:39.041 [2024-12-07 05:35:12.194348] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:39.041 [2024-12-07 05:35:12.194450] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:39.041 [2024-12-07 05:35:12.194509] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:39.041 pt1 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:39.041 "name": "raid_bdev1", 00:07:39.041 "uuid": "ad9f0507-a8fe-46a8-9da8-600ef57a8e24", 00:07:39.041 "strip_size_kb": 0, 00:07:39.041 "state": "configuring", 00:07:39.041 "raid_level": "raid1", 00:07:39.041 "superblock": true, 00:07:39.041 "num_base_bdevs": 2, 00:07:39.041 "num_base_bdevs_discovered": 1, 00:07:39.041 "num_base_bdevs_operational": 2, 00:07:39.041 "base_bdevs_list": [ 00:07:39.041 { 00:07:39.041 "name": "pt1", 00:07:39.041 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:39.041 "is_configured": true, 00:07:39.041 "data_offset": 2048, 00:07:39.041 "data_size": 63488 00:07:39.041 }, 00:07:39.041 { 00:07:39.041 "name": null, 00:07:39.041 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:39.041 "is_configured": false, 00:07:39.041 "data_offset": 2048, 00:07:39.041 "data_size": 63488 00:07:39.041 } 00:07:39.041 ] 00:07:39.041 }' 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:39.041 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.314 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:39.314 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:39.314 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:39.314 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:39.314 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.314 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.314 [2024-12-07 05:35:12.599367] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:39.314 [2024-12-07 05:35:12.599432] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:39.314 [2024-12-07 05:35:12.599456] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:39.314 [2024-12-07 05:35:12.599465] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:39.314 [2024-12-07 05:35:12.599890] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:39.314 [2024-12-07 05:35:12.599909] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:39.314 [2024-12-07 05:35:12.599986] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:39.314 [2024-12-07 05:35:12.600008] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:39.314 [2024-12-07 05:35:12.600099] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:39.314 [2024-12-07 05:35:12.600107] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:39.314 [2024-12-07 05:35:12.600343] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:39.314 [2024-12-07 05:35:12.600453] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:39.314 [2024-12-07 05:35:12.600466] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:39.314 [2024-12-07 05:35:12.600570] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:39.314 pt2 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:39.315 "name": "raid_bdev1", 00:07:39.315 "uuid": "ad9f0507-a8fe-46a8-9da8-600ef57a8e24", 00:07:39.315 "strip_size_kb": 0, 00:07:39.315 "state": "online", 00:07:39.315 "raid_level": "raid1", 00:07:39.315 "superblock": true, 00:07:39.315 "num_base_bdevs": 2, 00:07:39.315 "num_base_bdevs_discovered": 2, 00:07:39.315 "num_base_bdevs_operational": 2, 00:07:39.315 "base_bdevs_list": [ 00:07:39.315 { 00:07:39.315 "name": "pt1", 00:07:39.315 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:39.315 "is_configured": true, 00:07:39.315 "data_offset": 2048, 00:07:39.315 "data_size": 63488 00:07:39.315 }, 00:07:39.315 { 00:07:39.315 "name": "pt2", 00:07:39.315 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:39.315 "is_configured": true, 00:07:39.315 "data_offset": 2048, 00:07:39.315 "data_size": 63488 00:07:39.315 } 00:07:39.315 ] 00:07:39.315 }' 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:39.315 05:35:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.885 [2024-12-07 05:35:13.030907] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:39.885 "name": "raid_bdev1", 00:07:39.885 "aliases": [ 00:07:39.885 "ad9f0507-a8fe-46a8-9da8-600ef57a8e24" 00:07:39.885 ], 00:07:39.885 "product_name": "Raid Volume", 00:07:39.885 "block_size": 512, 00:07:39.885 "num_blocks": 63488, 00:07:39.885 "uuid": "ad9f0507-a8fe-46a8-9da8-600ef57a8e24", 00:07:39.885 "assigned_rate_limits": { 00:07:39.885 "rw_ios_per_sec": 0, 00:07:39.885 "rw_mbytes_per_sec": 0, 00:07:39.885 "r_mbytes_per_sec": 0, 00:07:39.885 "w_mbytes_per_sec": 0 00:07:39.885 }, 00:07:39.885 "claimed": false, 00:07:39.885 "zoned": false, 00:07:39.885 "supported_io_types": { 00:07:39.885 "read": true, 00:07:39.885 "write": true, 00:07:39.885 "unmap": false, 00:07:39.885 "flush": false, 00:07:39.885 "reset": true, 00:07:39.885 "nvme_admin": false, 00:07:39.885 "nvme_io": false, 00:07:39.885 "nvme_io_md": false, 00:07:39.885 "write_zeroes": true, 00:07:39.885 "zcopy": false, 00:07:39.885 "get_zone_info": false, 00:07:39.885 "zone_management": false, 00:07:39.885 "zone_append": false, 00:07:39.885 "compare": false, 00:07:39.885 "compare_and_write": false, 00:07:39.885 "abort": false, 00:07:39.885 "seek_hole": false, 00:07:39.885 "seek_data": false, 00:07:39.885 "copy": false, 00:07:39.885 "nvme_iov_md": false 00:07:39.885 }, 00:07:39.885 "memory_domains": [ 00:07:39.885 { 00:07:39.885 "dma_device_id": "system", 00:07:39.885 "dma_device_type": 1 00:07:39.885 }, 00:07:39.885 { 00:07:39.885 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.885 "dma_device_type": 2 00:07:39.885 }, 00:07:39.885 { 00:07:39.885 "dma_device_id": "system", 00:07:39.885 "dma_device_type": 1 00:07:39.885 }, 00:07:39.885 { 00:07:39.885 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.885 "dma_device_type": 2 00:07:39.885 } 00:07:39.885 ], 00:07:39.885 "driver_specific": { 00:07:39.885 "raid": { 00:07:39.885 "uuid": "ad9f0507-a8fe-46a8-9da8-600ef57a8e24", 00:07:39.885 "strip_size_kb": 0, 00:07:39.885 "state": "online", 00:07:39.885 "raid_level": "raid1", 00:07:39.885 "superblock": true, 00:07:39.885 "num_base_bdevs": 2, 00:07:39.885 "num_base_bdevs_discovered": 2, 00:07:39.885 "num_base_bdevs_operational": 2, 00:07:39.885 "base_bdevs_list": [ 00:07:39.885 { 00:07:39.885 "name": "pt1", 00:07:39.885 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:39.885 "is_configured": true, 00:07:39.885 "data_offset": 2048, 00:07:39.885 "data_size": 63488 00:07:39.885 }, 00:07:39.885 { 00:07:39.885 "name": "pt2", 00:07:39.885 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:39.885 "is_configured": true, 00:07:39.885 "data_offset": 2048, 00:07:39.885 "data_size": 63488 00:07:39.885 } 00:07:39.885 ] 00:07:39.885 } 00:07:39.885 } 00:07:39.885 }' 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:39.885 pt2' 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.885 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.145 [2024-12-07 05:35:13.282426] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' ad9f0507-a8fe-46a8-9da8-600ef57a8e24 '!=' ad9f0507-a8fe-46a8-9da8-600ef57a8e24 ']' 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.145 [2024-12-07 05:35:13.330112] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.145 "name": "raid_bdev1", 00:07:40.145 "uuid": "ad9f0507-a8fe-46a8-9da8-600ef57a8e24", 00:07:40.145 "strip_size_kb": 0, 00:07:40.145 "state": "online", 00:07:40.145 "raid_level": "raid1", 00:07:40.145 "superblock": true, 00:07:40.145 "num_base_bdevs": 2, 00:07:40.145 "num_base_bdevs_discovered": 1, 00:07:40.145 "num_base_bdevs_operational": 1, 00:07:40.145 "base_bdevs_list": [ 00:07:40.145 { 00:07:40.145 "name": null, 00:07:40.145 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:40.145 "is_configured": false, 00:07:40.145 "data_offset": 0, 00:07:40.145 "data_size": 63488 00:07:40.145 }, 00:07:40.145 { 00:07:40.145 "name": "pt2", 00:07:40.145 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:40.145 "is_configured": true, 00:07:40.145 "data_offset": 2048, 00:07:40.145 "data_size": 63488 00:07:40.145 } 00:07:40.145 ] 00:07:40.145 }' 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.145 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.404 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:40.404 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.404 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.404 [2024-12-07 05:35:13.753354] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:40.404 [2024-12-07 05:35:13.753425] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:40.404 [2024-12-07 05:35:13.753520] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:40.404 [2024-12-07 05:35:13.753569] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:40.404 [2024-12-07 05:35:13.753578] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:40.404 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.404 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.404 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:07:40.404 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.404 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.404 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.663 [2024-12-07 05:35:13.825244] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:40.663 [2024-12-07 05:35:13.825345] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:40.663 [2024-12-07 05:35:13.825382] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:07:40.663 [2024-12-07 05:35:13.825458] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:40.663 [2024-12-07 05:35:13.827649] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:40.663 [2024-12-07 05:35:13.827716] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:40.663 [2024-12-07 05:35:13.827809] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:40.663 [2024-12-07 05:35:13.827857] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:40.663 [2024-12-07 05:35:13.827983] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:07:40.663 [2024-12-07 05:35:13.828018] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:40.663 [2024-12-07 05:35:13.828278] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:40.663 [2024-12-07 05:35:13.828425] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:07:40.663 [2024-12-07 05:35:13.828464] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:07:40.663 [2024-12-07 05:35:13.828603] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:40.663 pt2 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.663 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.663 "name": "raid_bdev1", 00:07:40.664 "uuid": "ad9f0507-a8fe-46a8-9da8-600ef57a8e24", 00:07:40.664 "strip_size_kb": 0, 00:07:40.664 "state": "online", 00:07:40.664 "raid_level": "raid1", 00:07:40.664 "superblock": true, 00:07:40.664 "num_base_bdevs": 2, 00:07:40.664 "num_base_bdevs_discovered": 1, 00:07:40.664 "num_base_bdevs_operational": 1, 00:07:40.664 "base_bdevs_list": [ 00:07:40.664 { 00:07:40.664 "name": null, 00:07:40.664 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:40.664 "is_configured": false, 00:07:40.664 "data_offset": 2048, 00:07:40.664 "data_size": 63488 00:07:40.664 }, 00:07:40.664 { 00:07:40.664 "name": "pt2", 00:07:40.664 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:40.664 "is_configured": true, 00:07:40.664 "data_offset": 2048, 00:07:40.664 "data_size": 63488 00:07:40.664 } 00:07:40.664 ] 00:07:40.664 }' 00:07:40.664 05:35:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.664 05:35:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.923 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:40.923 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.923 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.923 [2024-12-07 05:35:14.284439] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:40.923 [2024-12-07 05:35:14.284464] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:40.923 [2024-12-07 05:35:14.284533] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:40.923 [2024-12-07 05:35:14.284577] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:40.923 [2024-12-07 05:35:14.284588] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:07:40.923 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.182 [2024-12-07 05:35:14.348314] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:41.182 [2024-12-07 05:35:14.348377] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:41.182 [2024-12-07 05:35:14.348398] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:07:41.182 [2024-12-07 05:35:14.348410] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:41.182 [2024-12-07 05:35:14.350602] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:41.182 [2024-12-07 05:35:14.350646] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:41.182 [2024-12-07 05:35:14.350719] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:41.182 [2024-12-07 05:35:14.350755] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:41.182 [2024-12-07 05:35:14.350853] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:07:41.182 [2024-12-07 05:35:14.350867] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:41.182 [2024-12-07 05:35:14.350892] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:07:41.182 [2024-12-07 05:35:14.350918] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:41.182 [2024-12-07 05:35:14.350985] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:07:41.182 [2024-12-07 05:35:14.350997] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:41.182 [2024-12-07 05:35:14.351227] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:07:41.182 [2024-12-07 05:35:14.351352] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:07:41.182 [2024-12-07 05:35:14.351360] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:07:41.182 [2024-12-07 05:35:14.351471] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:41.182 pt1 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:41.182 "name": "raid_bdev1", 00:07:41.182 "uuid": "ad9f0507-a8fe-46a8-9da8-600ef57a8e24", 00:07:41.182 "strip_size_kb": 0, 00:07:41.182 "state": "online", 00:07:41.182 "raid_level": "raid1", 00:07:41.182 "superblock": true, 00:07:41.182 "num_base_bdevs": 2, 00:07:41.182 "num_base_bdevs_discovered": 1, 00:07:41.182 "num_base_bdevs_operational": 1, 00:07:41.182 "base_bdevs_list": [ 00:07:41.182 { 00:07:41.182 "name": null, 00:07:41.182 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:41.182 "is_configured": false, 00:07:41.182 "data_offset": 2048, 00:07:41.182 "data_size": 63488 00:07:41.182 }, 00:07:41.182 { 00:07:41.182 "name": "pt2", 00:07:41.182 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:41.182 "is_configured": true, 00:07:41.182 "data_offset": 2048, 00:07:41.182 "data_size": 63488 00:07:41.182 } 00:07:41.182 ] 00:07:41.182 }' 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:41.182 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.441 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:41.441 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:07:41.441 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.441 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.441 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.441 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:07:41.441 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:41.441 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.441 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.441 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:07:41.441 [2024-12-07 05:35:14.803761] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' ad9f0507-a8fe-46a8-9da8-600ef57a8e24 '!=' ad9f0507-a8fe-46a8-9da8-600ef57a8e24 ']' 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 74144 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 74144 ']' 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 74144 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74144 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74144' 00:07:41.700 killing process with pid 74144 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 74144 00:07:41.700 [2024-12-07 05:35:14.864842] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:41.700 [2024-12-07 05:35:14.864913] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:41.700 [2024-12-07 05:35:14.864959] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:41.700 [2024-12-07 05:35:14.864967] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:07:41.700 05:35:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 74144 00:07:41.700 [2024-12-07 05:35:14.887502] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:41.958 05:35:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:41.958 00:07:41.958 real 0m4.734s 00:07:41.958 user 0m7.791s 00:07:41.958 sys 0m0.971s 00:07:41.958 05:35:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:41.958 ************************************ 00:07:41.958 END TEST raid_superblock_test 00:07:41.958 ************************************ 00:07:41.958 05:35:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.958 05:35:15 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:07:41.958 05:35:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:41.958 05:35:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:41.958 05:35:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:41.958 ************************************ 00:07:41.958 START TEST raid_read_error_test 00:07:41.958 ************************************ 00:07:41.958 05:35:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 read 00:07:41.958 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:07:41.958 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:41.958 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:41.958 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:41.958 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:41.958 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:41.958 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:41.958 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:41.958 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.62dfd1eSlw 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74462 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74462 00:07:41.959 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 74462 ']' 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:41.959 05:35:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.959 [2024-12-07 05:35:15.243384] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:41.959 [2024-12-07 05:35:15.243506] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74462 ] 00:07:42.218 [2024-12-07 05:35:15.378523] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:42.218 [2024-12-07 05:35:15.402798] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:42.218 [2024-12-07 05:35:15.444298] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:42.218 [2024-12-07 05:35:15.444342] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.786 BaseBdev1_malloc 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.786 true 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.786 [2024-12-07 05:35:16.103056] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:42.786 [2024-12-07 05:35:16.103169] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:42.786 [2024-12-07 05:35:16.103201] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:42.786 [2024-12-07 05:35:16.103210] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:42.786 [2024-12-07 05:35:16.105347] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:42.786 [2024-12-07 05:35:16.105384] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:42.786 BaseBdev1 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.786 BaseBdev2_malloc 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.786 true 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.786 [2024-12-07 05:35:16.143268] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:42.786 [2024-12-07 05:35:16.143317] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:42.786 [2024-12-07 05:35:16.143349] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:42.786 [2024-12-07 05:35:16.143365] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:42.786 [2024-12-07 05:35:16.145444] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:42.786 [2024-12-07 05:35:16.145479] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:42.786 BaseBdev2 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:42.786 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.045 [2024-12-07 05:35:16.155285] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:43.045 [2024-12-07 05:35:16.157154] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:43.045 [2024-12-07 05:35:16.157344] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:43.045 [2024-12-07 05:35:16.157360] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:43.045 [2024-12-07 05:35:16.157591] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:43.045 [2024-12-07 05:35:16.157785] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:43.045 [2024-12-07 05:35:16.157802] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:43.045 [2024-12-07 05:35:16.157921] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:43.045 "name": "raid_bdev1", 00:07:43.045 "uuid": "578b5e9a-f569-49cf-8e32-d03a9f2c89bc", 00:07:43.045 "strip_size_kb": 0, 00:07:43.045 "state": "online", 00:07:43.045 "raid_level": "raid1", 00:07:43.045 "superblock": true, 00:07:43.045 "num_base_bdevs": 2, 00:07:43.045 "num_base_bdevs_discovered": 2, 00:07:43.045 "num_base_bdevs_operational": 2, 00:07:43.045 "base_bdevs_list": [ 00:07:43.045 { 00:07:43.045 "name": "BaseBdev1", 00:07:43.045 "uuid": "c51d19c3-65cc-5a9e-9388-30bd255adcad", 00:07:43.045 "is_configured": true, 00:07:43.045 "data_offset": 2048, 00:07:43.045 "data_size": 63488 00:07:43.045 }, 00:07:43.045 { 00:07:43.045 "name": "BaseBdev2", 00:07:43.045 "uuid": "829c7c50-6698-5430-8415-b192d971619d", 00:07:43.045 "is_configured": true, 00:07:43.045 "data_offset": 2048, 00:07:43.045 "data_size": 63488 00:07:43.045 } 00:07:43.045 ] 00:07:43.045 }' 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:43.045 05:35:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.304 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:43.304 05:35:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:43.563 [2024-12-07 05:35:16.686769] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.501 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:44.501 "name": "raid_bdev1", 00:07:44.501 "uuid": "578b5e9a-f569-49cf-8e32-d03a9f2c89bc", 00:07:44.501 "strip_size_kb": 0, 00:07:44.501 "state": "online", 00:07:44.501 "raid_level": "raid1", 00:07:44.501 "superblock": true, 00:07:44.501 "num_base_bdevs": 2, 00:07:44.501 "num_base_bdevs_discovered": 2, 00:07:44.501 "num_base_bdevs_operational": 2, 00:07:44.501 "base_bdevs_list": [ 00:07:44.501 { 00:07:44.502 "name": "BaseBdev1", 00:07:44.502 "uuid": "c51d19c3-65cc-5a9e-9388-30bd255adcad", 00:07:44.502 "is_configured": true, 00:07:44.502 "data_offset": 2048, 00:07:44.502 "data_size": 63488 00:07:44.502 }, 00:07:44.502 { 00:07:44.502 "name": "BaseBdev2", 00:07:44.502 "uuid": "829c7c50-6698-5430-8415-b192d971619d", 00:07:44.502 "is_configured": true, 00:07:44.502 "data_offset": 2048, 00:07:44.502 "data_size": 63488 00:07:44.502 } 00:07:44.502 ] 00:07:44.502 }' 00:07:44.502 05:35:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:44.502 05:35:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.760 05:35:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:44.760 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.760 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.760 [2024-12-07 05:35:18.018132] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:44.761 [2024-12-07 05:35:18.018165] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:44.761 [2024-12-07 05:35:18.020825] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:44.761 [2024-12-07 05:35:18.020913] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:44.761 [2024-12-07 05:35:18.021023] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:44.761 [2024-12-07 05:35:18.021067] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:44.761 { 00:07:44.761 "results": [ 00:07:44.761 { 00:07:44.761 "job": "raid_bdev1", 00:07:44.761 "core_mask": "0x1", 00:07:44.761 "workload": "randrw", 00:07:44.761 "percentage": 50, 00:07:44.761 "status": "finished", 00:07:44.761 "queue_depth": 1, 00:07:44.761 "io_size": 131072, 00:07:44.761 "runtime": 1.332201, 00:07:44.761 "iops": 19385.212892048574, 00:07:44.761 "mibps": 2423.1516115060717, 00:07:44.761 "io_failed": 0, 00:07:44.761 "io_timeout": 0, 00:07:44.761 "avg_latency_us": 49.00280257189599, 00:07:44.761 "min_latency_us": 22.246288209606988, 00:07:44.761 "max_latency_us": 1409.4532751091704 00:07:44.761 } 00:07:44.761 ], 00:07:44.761 "core_count": 1 00:07:44.761 } 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74462 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 74462 ']' 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 74462 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74462 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74462' 00:07:44.761 killing process with pid 74462 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 74462 00:07:44.761 [2024-12-07 05:35:18.054600] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:44.761 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 74462 00:07:44.761 [2024-12-07 05:35:18.069982] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:45.020 05:35:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:45.020 05:35:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.62dfd1eSlw 00:07:45.020 05:35:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:45.020 05:35:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:07:45.020 ************************************ 00:07:45.020 END TEST raid_read_error_test 00:07:45.020 ************************************ 00:07:45.020 05:35:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:07:45.020 05:35:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:45.020 05:35:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:45.020 05:35:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:07:45.020 00:07:45.020 real 0m3.137s 00:07:45.020 user 0m4.010s 00:07:45.020 sys 0m0.445s 00:07:45.020 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:45.020 05:35:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.020 05:35:18 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:07:45.020 05:35:18 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:45.020 05:35:18 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:45.020 05:35:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:45.020 ************************************ 00:07:45.020 START TEST raid_write_error_test 00:07:45.020 ************************************ 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 write 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.RWi4T5fdZo 00:07:45.020 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74591 00:07:45.021 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:45.021 05:35:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74591 00:07:45.021 05:35:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 74591 ']' 00:07:45.021 05:35:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:45.021 05:35:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:45.021 05:35:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:45.021 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:45.021 05:35:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:45.021 05:35:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.280 [2024-12-07 05:35:18.438856] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:45.280 [2024-12-07 05:35:18.438973] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74591 ] 00:07:45.280 [2024-12-07 05:35:18.593309] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:45.280 [2024-12-07 05:35:18.617902] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:45.540 [2024-12-07 05:35:18.659664] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:45.540 [2024-12-07 05:35:18.659698] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:46.110 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:46.110 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.111 BaseBdev1_malloc 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.111 true 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.111 [2024-12-07 05:35:19.302655] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:46.111 [2024-12-07 05:35:19.302707] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:46.111 [2024-12-07 05:35:19.302732] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:46.111 [2024-12-07 05:35:19.302746] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:46.111 [2024-12-07 05:35:19.304836] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:46.111 [2024-12-07 05:35:19.304873] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:46.111 BaseBdev1 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.111 BaseBdev2_malloc 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.111 true 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.111 [2024-12-07 05:35:19.342995] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:46.111 [2024-12-07 05:35:19.343044] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:46.111 [2024-12-07 05:35:19.343077] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:46.111 [2024-12-07 05:35:19.343094] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:46.111 [2024-12-07 05:35:19.345208] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:46.111 [2024-12-07 05:35:19.345247] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:46.111 BaseBdev2 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.111 [2024-12-07 05:35:19.355013] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:46.111 [2024-12-07 05:35:19.356868] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:46.111 [2024-12-07 05:35:19.357049] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:46.111 [2024-12-07 05:35:19.357063] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:46.111 [2024-12-07 05:35:19.357315] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:46.111 [2024-12-07 05:35:19.357484] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:46.111 [2024-12-07 05:35:19.357496] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:46.111 [2024-12-07 05:35:19.357617] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.111 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:46.111 "name": "raid_bdev1", 00:07:46.111 "uuid": "30212b2b-6964-4927-95d5-74d4a8b3decc", 00:07:46.111 "strip_size_kb": 0, 00:07:46.111 "state": "online", 00:07:46.111 "raid_level": "raid1", 00:07:46.111 "superblock": true, 00:07:46.111 "num_base_bdevs": 2, 00:07:46.111 "num_base_bdevs_discovered": 2, 00:07:46.111 "num_base_bdevs_operational": 2, 00:07:46.111 "base_bdevs_list": [ 00:07:46.111 { 00:07:46.111 "name": "BaseBdev1", 00:07:46.111 "uuid": "6742d8d2-6d51-59bf-89f1-9f83fe601f08", 00:07:46.111 "is_configured": true, 00:07:46.111 "data_offset": 2048, 00:07:46.111 "data_size": 63488 00:07:46.111 }, 00:07:46.111 { 00:07:46.111 "name": "BaseBdev2", 00:07:46.112 "uuid": "29ac4a1f-e4a7-5e9e-bb30-8ee691894836", 00:07:46.112 "is_configured": true, 00:07:46.112 "data_offset": 2048, 00:07:46.112 "data_size": 63488 00:07:46.112 } 00:07:46.112 ] 00:07:46.112 }' 00:07:46.112 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:46.112 05:35:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.681 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:46.681 05:35:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:46.681 [2024-12-07 05:35:19.858589] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.630 [2024-12-07 05:35:20.778508] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:07:47.630 [2024-12-07 05:35:20.778567] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:47.630 [2024-12-07 05:35:20.778800] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002a10 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:47.630 "name": "raid_bdev1", 00:07:47.630 "uuid": "30212b2b-6964-4927-95d5-74d4a8b3decc", 00:07:47.630 "strip_size_kb": 0, 00:07:47.630 "state": "online", 00:07:47.630 "raid_level": "raid1", 00:07:47.630 "superblock": true, 00:07:47.630 "num_base_bdevs": 2, 00:07:47.630 "num_base_bdevs_discovered": 1, 00:07:47.630 "num_base_bdevs_operational": 1, 00:07:47.630 "base_bdevs_list": [ 00:07:47.630 { 00:07:47.630 "name": null, 00:07:47.630 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:47.630 "is_configured": false, 00:07:47.630 "data_offset": 0, 00:07:47.630 "data_size": 63488 00:07:47.630 }, 00:07:47.630 { 00:07:47.630 "name": "BaseBdev2", 00:07:47.630 "uuid": "29ac4a1f-e4a7-5e9e-bb30-8ee691894836", 00:07:47.630 "is_configured": true, 00:07:47.630 "data_offset": 2048, 00:07:47.630 "data_size": 63488 00:07:47.630 } 00:07:47.630 ] 00:07:47.630 }' 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:47.630 05:35:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.890 05:35:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:47.890 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:47.890 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.890 [2024-12-07 05:35:21.239287] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:47.890 [2024-12-07 05:35:21.239390] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:47.890 [2024-12-07 05:35:21.241877] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:47.890 [2024-12-07 05:35:21.241969] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:47.890 [2024-12-07 05:35:21.242068] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:47.890 [2024-12-07 05:35:21.242115] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:47.890 { 00:07:47.890 "results": [ 00:07:47.890 { 00:07:47.890 "job": "raid_bdev1", 00:07:47.890 "core_mask": "0x1", 00:07:47.890 "workload": "randrw", 00:07:47.890 "percentage": 50, 00:07:47.890 "status": "finished", 00:07:47.890 "queue_depth": 1, 00:07:47.890 "io_size": 131072, 00:07:47.890 "runtime": 1.381561, 00:07:47.890 "iops": 23260.645023998215, 00:07:47.890 "mibps": 2907.580627999777, 00:07:47.890 "io_failed": 0, 00:07:47.890 "io_timeout": 0, 00:07:47.890 "avg_latency_us": 40.34917626289145, 00:07:47.890 "min_latency_us": 22.134497816593885, 00:07:47.890 "max_latency_us": 1345.0620087336245 00:07:47.890 } 00:07:47.890 ], 00:07:47.890 "core_count": 1 00:07:47.890 } 00:07:47.890 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:47.890 05:35:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74591 00:07:47.890 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 74591 ']' 00:07:47.890 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 74591 00:07:47.890 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:07:47.890 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:47.890 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74591 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74591' 00:07:48.152 killing process with pid 74591 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 74591 00:07:48.152 [2024-12-07 05:35:21.273693] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 74591 00:07:48.152 [2024-12-07 05:35:21.288422] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.RWi4T5fdZo 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:48.152 ************************************ 00:07:48.152 END TEST raid_write_error_test 00:07:48.152 ************************************ 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:07:48.152 00:07:48.152 real 0m3.149s 00:07:48.152 user 0m4.027s 00:07:48.152 sys 0m0.455s 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:48.152 05:35:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.414 05:35:21 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:07:48.414 05:35:21 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:48.414 05:35:21 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:07:48.414 05:35:21 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:48.414 05:35:21 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:48.414 05:35:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:48.414 ************************************ 00:07:48.414 START TEST raid_state_function_test 00:07:48.414 ************************************ 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 false 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=74718 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74718' 00:07:48.414 Process raid pid: 74718 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 74718 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 74718 ']' 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:48.414 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:48.414 05:35:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.414 [2024-12-07 05:35:21.652668] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:48.414 [2024-12-07 05:35:21.652866] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:48.674 [2024-12-07 05:35:21.809017] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:48.674 [2024-12-07 05:35:21.833475] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:48.674 [2024-12-07 05:35:21.874895] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:48.674 [2024-12-07 05:35:21.874934] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.246 [2024-12-07 05:35:22.501007] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:49.246 [2024-12-07 05:35:22.501066] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:49.246 [2024-12-07 05:35:22.501077] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:49.246 [2024-12-07 05:35:22.501087] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:49.246 [2024-12-07 05:35:22.501093] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:49.246 [2024-12-07 05:35:22.501105] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:49.246 "name": "Existed_Raid", 00:07:49.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:49.246 "strip_size_kb": 64, 00:07:49.246 "state": "configuring", 00:07:49.246 "raid_level": "raid0", 00:07:49.246 "superblock": false, 00:07:49.246 "num_base_bdevs": 3, 00:07:49.246 "num_base_bdevs_discovered": 0, 00:07:49.246 "num_base_bdevs_operational": 3, 00:07:49.246 "base_bdevs_list": [ 00:07:49.246 { 00:07:49.246 "name": "BaseBdev1", 00:07:49.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:49.246 "is_configured": false, 00:07:49.246 "data_offset": 0, 00:07:49.246 "data_size": 0 00:07:49.246 }, 00:07:49.246 { 00:07:49.246 "name": "BaseBdev2", 00:07:49.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:49.246 "is_configured": false, 00:07:49.246 "data_offset": 0, 00:07:49.246 "data_size": 0 00:07:49.246 }, 00:07:49.246 { 00:07:49.246 "name": "BaseBdev3", 00:07:49.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:49.246 "is_configured": false, 00:07:49.246 "data_offset": 0, 00:07:49.246 "data_size": 0 00:07:49.246 } 00:07:49.246 ] 00:07:49.246 }' 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:49.246 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.816 [2024-12-07 05:35:22.892248] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:49.816 [2024-12-07 05:35:22.892334] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.816 [2024-12-07 05:35:22.904249] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:49.816 [2024-12-07 05:35:22.904345] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:49.816 [2024-12-07 05:35:22.904373] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:49.816 [2024-12-07 05:35:22.904395] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:49.816 [2024-12-07 05:35:22.904412] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:49.816 [2024-12-07 05:35:22.904432] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.816 [2024-12-07 05:35:22.924887] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:49.816 BaseBdev1 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.816 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.816 [ 00:07:49.816 { 00:07:49.816 "name": "BaseBdev1", 00:07:49.816 "aliases": [ 00:07:49.816 "53788d16-f427-4437-9ec9-c6ca6eaa544f" 00:07:49.816 ], 00:07:49.816 "product_name": "Malloc disk", 00:07:49.817 "block_size": 512, 00:07:49.817 "num_blocks": 65536, 00:07:49.817 "uuid": "53788d16-f427-4437-9ec9-c6ca6eaa544f", 00:07:49.817 "assigned_rate_limits": { 00:07:49.817 "rw_ios_per_sec": 0, 00:07:49.817 "rw_mbytes_per_sec": 0, 00:07:49.817 "r_mbytes_per_sec": 0, 00:07:49.817 "w_mbytes_per_sec": 0 00:07:49.817 }, 00:07:49.817 "claimed": true, 00:07:49.817 "claim_type": "exclusive_write", 00:07:49.817 "zoned": false, 00:07:49.817 "supported_io_types": { 00:07:49.817 "read": true, 00:07:49.817 "write": true, 00:07:49.817 "unmap": true, 00:07:49.817 "flush": true, 00:07:49.817 "reset": true, 00:07:49.817 "nvme_admin": false, 00:07:49.817 "nvme_io": false, 00:07:49.817 "nvme_io_md": false, 00:07:49.817 "write_zeroes": true, 00:07:49.817 "zcopy": true, 00:07:49.817 "get_zone_info": false, 00:07:49.817 "zone_management": false, 00:07:49.817 "zone_append": false, 00:07:49.817 "compare": false, 00:07:49.817 "compare_and_write": false, 00:07:49.817 "abort": true, 00:07:49.817 "seek_hole": false, 00:07:49.817 "seek_data": false, 00:07:49.817 "copy": true, 00:07:49.817 "nvme_iov_md": false 00:07:49.817 }, 00:07:49.817 "memory_domains": [ 00:07:49.817 { 00:07:49.817 "dma_device_id": "system", 00:07:49.817 "dma_device_type": 1 00:07:49.817 }, 00:07:49.817 { 00:07:49.817 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:49.817 "dma_device_type": 2 00:07:49.817 } 00:07:49.817 ], 00:07:49.817 "driver_specific": {} 00:07:49.817 } 00:07:49.817 ] 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.817 05:35:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:49.817 "name": "Existed_Raid", 00:07:49.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:49.817 "strip_size_kb": 64, 00:07:49.817 "state": "configuring", 00:07:49.817 "raid_level": "raid0", 00:07:49.817 "superblock": false, 00:07:49.817 "num_base_bdevs": 3, 00:07:49.817 "num_base_bdevs_discovered": 1, 00:07:49.817 "num_base_bdevs_operational": 3, 00:07:49.817 "base_bdevs_list": [ 00:07:49.817 { 00:07:49.817 "name": "BaseBdev1", 00:07:49.817 "uuid": "53788d16-f427-4437-9ec9-c6ca6eaa544f", 00:07:49.817 "is_configured": true, 00:07:49.817 "data_offset": 0, 00:07:49.817 "data_size": 65536 00:07:49.817 }, 00:07:49.817 { 00:07:49.817 "name": "BaseBdev2", 00:07:49.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:49.817 "is_configured": false, 00:07:49.817 "data_offset": 0, 00:07:49.817 "data_size": 0 00:07:49.817 }, 00:07:49.817 { 00:07:49.817 "name": "BaseBdev3", 00:07:49.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:49.817 "is_configured": false, 00:07:49.817 "data_offset": 0, 00:07:49.817 "data_size": 0 00:07:49.817 } 00:07:49.817 ] 00:07:49.817 }' 00:07:49.817 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:49.817 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.078 [2024-12-07 05:35:23.352173] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:50.078 [2024-12-07 05:35:23.352257] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.078 [2024-12-07 05:35:23.360195] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:50.078 [2024-12-07 05:35:23.362035] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:50.078 [2024-12-07 05:35:23.362124] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:50.078 [2024-12-07 05:35:23.362138] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:50.078 [2024-12-07 05:35:23.362148] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.078 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:50.078 "name": "Existed_Raid", 00:07:50.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:50.078 "strip_size_kb": 64, 00:07:50.078 "state": "configuring", 00:07:50.078 "raid_level": "raid0", 00:07:50.078 "superblock": false, 00:07:50.078 "num_base_bdevs": 3, 00:07:50.078 "num_base_bdevs_discovered": 1, 00:07:50.078 "num_base_bdevs_operational": 3, 00:07:50.078 "base_bdevs_list": [ 00:07:50.078 { 00:07:50.078 "name": "BaseBdev1", 00:07:50.078 "uuid": "53788d16-f427-4437-9ec9-c6ca6eaa544f", 00:07:50.078 "is_configured": true, 00:07:50.078 "data_offset": 0, 00:07:50.078 "data_size": 65536 00:07:50.078 }, 00:07:50.078 { 00:07:50.078 "name": "BaseBdev2", 00:07:50.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:50.078 "is_configured": false, 00:07:50.078 "data_offset": 0, 00:07:50.078 "data_size": 0 00:07:50.078 }, 00:07:50.078 { 00:07:50.078 "name": "BaseBdev3", 00:07:50.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:50.078 "is_configured": false, 00:07:50.078 "data_offset": 0, 00:07:50.079 "data_size": 0 00:07:50.079 } 00:07:50.079 ] 00:07:50.079 }' 00:07:50.079 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:50.079 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.649 [2024-12-07 05:35:23.790329] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:50.649 BaseBdev2 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.649 [ 00:07:50.649 { 00:07:50.649 "name": "BaseBdev2", 00:07:50.649 "aliases": [ 00:07:50.649 "fb2a76d7-adce-4f79-8a10-90e3d3848cde" 00:07:50.649 ], 00:07:50.649 "product_name": "Malloc disk", 00:07:50.649 "block_size": 512, 00:07:50.649 "num_blocks": 65536, 00:07:50.649 "uuid": "fb2a76d7-adce-4f79-8a10-90e3d3848cde", 00:07:50.649 "assigned_rate_limits": { 00:07:50.649 "rw_ios_per_sec": 0, 00:07:50.649 "rw_mbytes_per_sec": 0, 00:07:50.649 "r_mbytes_per_sec": 0, 00:07:50.649 "w_mbytes_per_sec": 0 00:07:50.649 }, 00:07:50.649 "claimed": true, 00:07:50.649 "claim_type": "exclusive_write", 00:07:50.649 "zoned": false, 00:07:50.649 "supported_io_types": { 00:07:50.649 "read": true, 00:07:50.649 "write": true, 00:07:50.649 "unmap": true, 00:07:50.649 "flush": true, 00:07:50.649 "reset": true, 00:07:50.649 "nvme_admin": false, 00:07:50.649 "nvme_io": false, 00:07:50.649 "nvme_io_md": false, 00:07:50.649 "write_zeroes": true, 00:07:50.649 "zcopy": true, 00:07:50.649 "get_zone_info": false, 00:07:50.649 "zone_management": false, 00:07:50.649 "zone_append": false, 00:07:50.649 "compare": false, 00:07:50.649 "compare_and_write": false, 00:07:50.649 "abort": true, 00:07:50.649 "seek_hole": false, 00:07:50.649 "seek_data": false, 00:07:50.649 "copy": true, 00:07:50.649 "nvme_iov_md": false 00:07:50.649 }, 00:07:50.649 "memory_domains": [ 00:07:50.649 { 00:07:50.649 "dma_device_id": "system", 00:07:50.649 "dma_device_type": 1 00:07:50.649 }, 00:07:50.649 { 00:07:50.649 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:50.649 "dma_device_type": 2 00:07:50.649 } 00:07:50.649 ], 00:07:50.649 "driver_specific": {} 00:07:50.649 } 00:07:50.649 ] 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.649 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:50.649 "name": "Existed_Raid", 00:07:50.649 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:50.649 "strip_size_kb": 64, 00:07:50.649 "state": "configuring", 00:07:50.649 "raid_level": "raid0", 00:07:50.649 "superblock": false, 00:07:50.649 "num_base_bdevs": 3, 00:07:50.649 "num_base_bdevs_discovered": 2, 00:07:50.650 "num_base_bdevs_operational": 3, 00:07:50.650 "base_bdevs_list": [ 00:07:50.650 { 00:07:50.650 "name": "BaseBdev1", 00:07:50.650 "uuid": "53788d16-f427-4437-9ec9-c6ca6eaa544f", 00:07:50.650 "is_configured": true, 00:07:50.650 "data_offset": 0, 00:07:50.650 "data_size": 65536 00:07:50.650 }, 00:07:50.650 { 00:07:50.650 "name": "BaseBdev2", 00:07:50.650 "uuid": "fb2a76d7-adce-4f79-8a10-90e3d3848cde", 00:07:50.650 "is_configured": true, 00:07:50.650 "data_offset": 0, 00:07:50.650 "data_size": 65536 00:07:50.650 }, 00:07:50.650 { 00:07:50.650 "name": "BaseBdev3", 00:07:50.650 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:50.650 "is_configured": false, 00:07:50.650 "data_offset": 0, 00:07:50.650 "data_size": 0 00:07:50.650 } 00:07:50.650 ] 00:07:50.650 }' 00:07:50.650 05:35:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:50.650 05:35:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.911 [2024-12-07 05:35:24.243857] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:50.911 [2024-12-07 05:35:24.243956] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:50.911 [2024-12-07 05:35:24.244023] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:50.911 [2024-12-07 05:35:24.245068] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:50.911 [2024-12-07 05:35:24.245508] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:50.911 [2024-12-07 05:35:24.245573] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:50.911 [2024-12-07 05:35:24.246241] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:50.911 BaseBdev3 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.911 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.911 [ 00:07:50.911 { 00:07:50.911 "name": "BaseBdev3", 00:07:50.911 "aliases": [ 00:07:50.911 "9997bb6b-3ac7-4e0f-95ff-29a3a107e9d7" 00:07:50.911 ], 00:07:50.911 "product_name": "Malloc disk", 00:07:50.911 "block_size": 512, 00:07:50.911 "num_blocks": 65536, 00:07:50.911 "uuid": "9997bb6b-3ac7-4e0f-95ff-29a3a107e9d7", 00:07:50.911 "assigned_rate_limits": { 00:07:50.911 "rw_ios_per_sec": 0, 00:07:50.911 "rw_mbytes_per_sec": 0, 00:07:50.911 "r_mbytes_per_sec": 0, 00:07:50.911 "w_mbytes_per_sec": 0 00:07:51.172 }, 00:07:51.172 "claimed": true, 00:07:51.172 "claim_type": "exclusive_write", 00:07:51.172 "zoned": false, 00:07:51.172 "supported_io_types": { 00:07:51.172 "read": true, 00:07:51.172 "write": true, 00:07:51.172 "unmap": true, 00:07:51.172 "flush": true, 00:07:51.172 "reset": true, 00:07:51.172 "nvme_admin": false, 00:07:51.172 "nvme_io": false, 00:07:51.172 "nvme_io_md": false, 00:07:51.172 "write_zeroes": true, 00:07:51.172 "zcopy": true, 00:07:51.172 "get_zone_info": false, 00:07:51.172 "zone_management": false, 00:07:51.172 "zone_append": false, 00:07:51.172 "compare": false, 00:07:51.172 "compare_and_write": false, 00:07:51.172 "abort": true, 00:07:51.172 "seek_hole": false, 00:07:51.172 "seek_data": false, 00:07:51.172 "copy": true, 00:07:51.172 "nvme_iov_md": false 00:07:51.172 }, 00:07:51.172 "memory_domains": [ 00:07:51.172 { 00:07:51.172 "dma_device_id": "system", 00:07:51.172 "dma_device_type": 1 00:07:51.172 }, 00:07:51.172 { 00:07:51.172 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:51.172 "dma_device_type": 2 00:07:51.172 } 00:07:51.172 ], 00:07:51.172 "driver_specific": {} 00:07:51.172 } 00:07:51.172 ] 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:51.172 "name": "Existed_Raid", 00:07:51.172 "uuid": "066c4523-3620-44d4-85db-720f02bd3b44", 00:07:51.172 "strip_size_kb": 64, 00:07:51.172 "state": "online", 00:07:51.172 "raid_level": "raid0", 00:07:51.172 "superblock": false, 00:07:51.172 "num_base_bdevs": 3, 00:07:51.172 "num_base_bdevs_discovered": 3, 00:07:51.172 "num_base_bdevs_operational": 3, 00:07:51.172 "base_bdevs_list": [ 00:07:51.172 { 00:07:51.172 "name": "BaseBdev1", 00:07:51.172 "uuid": "53788d16-f427-4437-9ec9-c6ca6eaa544f", 00:07:51.172 "is_configured": true, 00:07:51.172 "data_offset": 0, 00:07:51.172 "data_size": 65536 00:07:51.172 }, 00:07:51.172 { 00:07:51.172 "name": "BaseBdev2", 00:07:51.172 "uuid": "fb2a76d7-adce-4f79-8a10-90e3d3848cde", 00:07:51.172 "is_configured": true, 00:07:51.172 "data_offset": 0, 00:07:51.172 "data_size": 65536 00:07:51.172 }, 00:07:51.172 { 00:07:51.172 "name": "BaseBdev3", 00:07:51.172 "uuid": "9997bb6b-3ac7-4e0f-95ff-29a3a107e9d7", 00:07:51.172 "is_configured": true, 00:07:51.172 "data_offset": 0, 00:07:51.172 "data_size": 65536 00:07:51.172 } 00:07:51.172 ] 00:07:51.172 }' 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:51.172 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.432 [2024-12-07 05:35:24.707303] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.432 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:51.432 "name": "Existed_Raid", 00:07:51.432 "aliases": [ 00:07:51.432 "066c4523-3620-44d4-85db-720f02bd3b44" 00:07:51.432 ], 00:07:51.432 "product_name": "Raid Volume", 00:07:51.432 "block_size": 512, 00:07:51.432 "num_blocks": 196608, 00:07:51.432 "uuid": "066c4523-3620-44d4-85db-720f02bd3b44", 00:07:51.432 "assigned_rate_limits": { 00:07:51.432 "rw_ios_per_sec": 0, 00:07:51.432 "rw_mbytes_per_sec": 0, 00:07:51.432 "r_mbytes_per_sec": 0, 00:07:51.432 "w_mbytes_per_sec": 0 00:07:51.432 }, 00:07:51.432 "claimed": false, 00:07:51.432 "zoned": false, 00:07:51.432 "supported_io_types": { 00:07:51.432 "read": true, 00:07:51.432 "write": true, 00:07:51.432 "unmap": true, 00:07:51.432 "flush": true, 00:07:51.432 "reset": true, 00:07:51.432 "nvme_admin": false, 00:07:51.432 "nvme_io": false, 00:07:51.432 "nvme_io_md": false, 00:07:51.432 "write_zeroes": true, 00:07:51.432 "zcopy": false, 00:07:51.432 "get_zone_info": false, 00:07:51.432 "zone_management": false, 00:07:51.432 "zone_append": false, 00:07:51.432 "compare": false, 00:07:51.432 "compare_and_write": false, 00:07:51.432 "abort": false, 00:07:51.432 "seek_hole": false, 00:07:51.432 "seek_data": false, 00:07:51.432 "copy": false, 00:07:51.432 "nvme_iov_md": false 00:07:51.432 }, 00:07:51.432 "memory_domains": [ 00:07:51.432 { 00:07:51.432 "dma_device_id": "system", 00:07:51.432 "dma_device_type": 1 00:07:51.432 }, 00:07:51.432 { 00:07:51.432 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:51.432 "dma_device_type": 2 00:07:51.432 }, 00:07:51.432 { 00:07:51.432 "dma_device_id": "system", 00:07:51.432 "dma_device_type": 1 00:07:51.432 }, 00:07:51.432 { 00:07:51.432 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:51.432 "dma_device_type": 2 00:07:51.432 }, 00:07:51.432 { 00:07:51.432 "dma_device_id": "system", 00:07:51.432 "dma_device_type": 1 00:07:51.432 }, 00:07:51.432 { 00:07:51.432 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:51.433 "dma_device_type": 2 00:07:51.433 } 00:07:51.433 ], 00:07:51.433 "driver_specific": { 00:07:51.433 "raid": { 00:07:51.433 "uuid": "066c4523-3620-44d4-85db-720f02bd3b44", 00:07:51.433 "strip_size_kb": 64, 00:07:51.433 "state": "online", 00:07:51.433 "raid_level": "raid0", 00:07:51.433 "superblock": false, 00:07:51.433 "num_base_bdevs": 3, 00:07:51.433 "num_base_bdevs_discovered": 3, 00:07:51.433 "num_base_bdevs_operational": 3, 00:07:51.433 "base_bdevs_list": [ 00:07:51.433 { 00:07:51.433 "name": "BaseBdev1", 00:07:51.433 "uuid": "53788d16-f427-4437-9ec9-c6ca6eaa544f", 00:07:51.433 "is_configured": true, 00:07:51.433 "data_offset": 0, 00:07:51.433 "data_size": 65536 00:07:51.433 }, 00:07:51.433 { 00:07:51.433 "name": "BaseBdev2", 00:07:51.433 "uuid": "fb2a76d7-adce-4f79-8a10-90e3d3848cde", 00:07:51.433 "is_configured": true, 00:07:51.433 "data_offset": 0, 00:07:51.433 "data_size": 65536 00:07:51.433 }, 00:07:51.433 { 00:07:51.433 "name": "BaseBdev3", 00:07:51.433 "uuid": "9997bb6b-3ac7-4e0f-95ff-29a3a107e9d7", 00:07:51.433 "is_configured": true, 00:07:51.433 "data_offset": 0, 00:07:51.433 "data_size": 65536 00:07:51.433 } 00:07:51.433 ] 00:07:51.433 } 00:07:51.433 } 00:07:51.433 }' 00:07:51.433 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:51.433 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:51.433 BaseBdev2 00:07:51.433 BaseBdev3' 00:07:51.433 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.693 [2024-12-07 05:35:24.986586] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:51.693 [2024-12-07 05:35:24.986616] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:51.693 [2024-12-07 05:35:24.986706] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:07:51.693 05:35:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.693 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:51.693 "name": "Existed_Raid", 00:07:51.693 "uuid": "066c4523-3620-44d4-85db-720f02bd3b44", 00:07:51.693 "strip_size_kb": 64, 00:07:51.693 "state": "offline", 00:07:51.693 "raid_level": "raid0", 00:07:51.693 "superblock": false, 00:07:51.693 "num_base_bdevs": 3, 00:07:51.693 "num_base_bdevs_discovered": 2, 00:07:51.694 "num_base_bdevs_operational": 2, 00:07:51.694 "base_bdevs_list": [ 00:07:51.694 { 00:07:51.694 "name": null, 00:07:51.694 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:51.694 "is_configured": false, 00:07:51.694 "data_offset": 0, 00:07:51.694 "data_size": 65536 00:07:51.694 }, 00:07:51.694 { 00:07:51.694 "name": "BaseBdev2", 00:07:51.694 "uuid": "fb2a76d7-adce-4f79-8a10-90e3d3848cde", 00:07:51.694 "is_configured": true, 00:07:51.694 "data_offset": 0, 00:07:51.694 "data_size": 65536 00:07:51.694 }, 00:07:51.694 { 00:07:51.694 "name": "BaseBdev3", 00:07:51.694 "uuid": "9997bb6b-3ac7-4e0f-95ff-29a3a107e9d7", 00:07:51.694 "is_configured": true, 00:07:51.694 "data_offset": 0, 00:07:51.694 "data_size": 65536 00:07:51.694 } 00:07:51.694 ] 00:07:51.694 }' 00:07:51.694 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:51.694 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.260 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:52.260 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:52.260 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.261 [2024-12-07 05:35:25.532866] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.261 [2024-12-07 05:35:25.603864] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:52.261 [2024-12-07 05:35:25.603974] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.261 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.520 BaseBdev2 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.520 [ 00:07:52.520 { 00:07:52.520 "name": "BaseBdev2", 00:07:52.520 "aliases": [ 00:07:52.520 "28903eab-c78e-4921-81b8-6458b54c5a6a" 00:07:52.520 ], 00:07:52.520 "product_name": "Malloc disk", 00:07:52.520 "block_size": 512, 00:07:52.520 "num_blocks": 65536, 00:07:52.520 "uuid": "28903eab-c78e-4921-81b8-6458b54c5a6a", 00:07:52.520 "assigned_rate_limits": { 00:07:52.520 "rw_ios_per_sec": 0, 00:07:52.520 "rw_mbytes_per_sec": 0, 00:07:52.520 "r_mbytes_per_sec": 0, 00:07:52.520 "w_mbytes_per_sec": 0 00:07:52.520 }, 00:07:52.520 "claimed": false, 00:07:52.520 "zoned": false, 00:07:52.520 "supported_io_types": { 00:07:52.520 "read": true, 00:07:52.520 "write": true, 00:07:52.520 "unmap": true, 00:07:52.520 "flush": true, 00:07:52.520 "reset": true, 00:07:52.520 "nvme_admin": false, 00:07:52.520 "nvme_io": false, 00:07:52.520 "nvme_io_md": false, 00:07:52.520 "write_zeroes": true, 00:07:52.520 "zcopy": true, 00:07:52.520 "get_zone_info": false, 00:07:52.520 "zone_management": false, 00:07:52.520 "zone_append": false, 00:07:52.520 "compare": false, 00:07:52.520 "compare_and_write": false, 00:07:52.520 "abort": true, 00:07:52.520 "seek_hole": false, 00:07:52.520 "seek_data": false, 00:07:52.520 "copy": true, 00:07:52.520 "nvme_iov_md": false 00:07:52.520 }, 00:07:52.520 "memory_domains": [ 00:07:52.520 { 00:07:52.520 "dma_device_id": "system", 00:07:52.520 "dma_device_type": 1 00:07:52.520 }, 00:07:52.520 { 00:07:52.520 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:52.520 "dma_device_type": 2 00:07:52.520 } 00:07:52.520 ], 00:07:52.520 "driver_specific": {} 00:07:52.520 } 00:07:52.520 ] 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.520 BaseBdev3 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.520 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.520 [ 00:07:52.520 { 00:07:52.520 "name": "BaseBdev3", 00:07:52.520 "aliases": [ 00:07:52.520 "0a91e3ae-7350-4acc-ac02-2f2f84f05b35" 00:07:52.520 ], 00:07:52.520 "product_name": "Malloc disk", 00:07:52.520 "block_size": 512, 00:07:52.520 "num_blocks": 65536, 00:07:52.520 "uuid": "0a91e3ae-7350-4acc-ac02-2f2f84f05b35", 00:07:52.520 "assigned_rate_limits": { 00:07:52.520 "rw_ios_per_sec": 0, 00:07:52.520 "rw_mbytes_per_sec": 0, 00:07:52.520 "r_mbytes_per_sec": 0, 00:07:52.520 "w_mbytes_per_sec": 0 00:07:52.520 }, 00:07:52.520 "claimed": false, 00:07:52.520 "zoned": false, 00:07:52.520 "supported_io_types": { 00:07:52.521 "read": true, 00:07:52.521 "write": true, 00:07:52.521 "unmap": true, 00:07:52.521 "flush": true, 00:07:52.521 "reset": true, 00:07:52.521 "nvme_admin": false, 00:07:52.521 "nvme_io": false, 00:07:52.521 "nvme_io_md": false, 00:07:52.521 "write_zeroes": true, 00:07:52.521 "zcopy": true, 00:07:52.521 "get_zone_info": false, 00:07:52.521 "zone_management": false, 00:07:52.521 "zone_append": false, 00:07:52.521 "compare": false, 00:07:52.521 "compare_and_write": false, 00:07:52.521 "abort": true, 00:07:52.521 "seek_hole": false, 00:07:52.521 "seek_data": false, 00:07:52.521 "copy": true, 00:07:52.521 "nvme_iov_md": false 00:07:52.521 }, 00:07:52.521 "memory_domains": [ 00:07:52.521 { 00:07:52.521 "dma_device_id": "system", 00:07:52.521 "dma_device_type": 1 00:07:52.521 }, 00:07:52.521 { 00:07:52.521 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:52.521 "dma_device_type": 2 00:07:52.521 } 00:07:52.521 ], 00:07:52.521 "driver_specific": {} 00:07:52.521 } 00:07:52.521 ] 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.521 [2024-12-07 05:35:25.763003] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:52.521 [2024-12-07 05:35:25.763108] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:52.521 [2024-12-07 05:35:25.763151] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:52.521 [2024-12-07 05:35:25.764975] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:52.521 "name": "Existed_Raid", 00:07:52.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.521 "strip_size_kb": 64, 00:07:52.521 "state": "configuring", 00:07:52.521 "raid_level": "raid0", 00:07:52.521 "superblock": false, 00:07:52.521 "num_base_bdevs": 3, 00:07:52.521 "num_base_bdevs_discovered": 2, 00:07:52.521 "num_base_bdevs_operational": 3, 00:07:52.521 "base_bdevs_list": [ 00:07:52.521 { 00:07:52.521 "name": "BaseBdev1", 00:07:52.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.521 "is_configured": false, 00:07:52.521 "data_offset": 0, 00:07:52.521 "data_size": 0 00:07:52.521 }, 00:07:52.521 { 00:07:52.521 "name": "BaseBdev2", 00:07:52.521 "uuid": "28903eab-c78e-4921-81b8-6458b54c5a6a", 00:07:52.521 "is_configured": true, 00:07:52.521 "data_offset": 0, 00:07:52.521 "data_size": 65536 00:07:52.521 }, 00:07:52.521 { 00:07:52.521 "name": "BaseBdev3", 00:07:52.521 "uuid": "0a91e3ae-7350-4acc-ac02-2f2f84f05b35", 00:07:52.521 "is_configured": true, 00:07:52.521 "data_offset": 0, 00:07:52.521 "data_size": 65536 00:07:52.521 } 00:07:52.521 ] 00:07:52.521 }' 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:52.521 05:35:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.101 [2024-12-07 05:35:26.210430] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.101 "name": "Existed_Raid", 00:07:53.101 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.101 "strip_size_kb": 64, 00:07:53.101 "state": "configuring", 00:07:53.101 "raid_level": "raid0", 00:07:53.101 "superblock": false, 00:07:53.101 "num_base_bdevs": 3, 00:07:53.101 "num_base_bdevs_discovered": 1, 00:07:53.101 "num_base_bdevs_operational": 3, 00:07:53.101 "base_bdevs_list": [ 00:07:53.101 { 00:07:53.101 "name": "BaseBdev1", 00:07:53.101 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.101 "is_configured": false, 00:07:53.101 "data_offset": 0, 00:07:53.101 "data_size": 0 00:07:53.101 }, 00:07:53.101 { 00:07:53.101 "name": null, 00:07:53.101 "uuid": "28903eab-c78e-4921-81b8-6458b54c5a6a", 00:07:53.101 "is_configured": false, 00:07:53.101 "data_offset": 0, 00:07:53.101 "data_size": 65536 00:07:53.101 }, 00:07:53.101 { 00:07:53.101 "name": "BaseBdev3", 00:07:53.101 "uuid": "0a91e3ae-7350-4acc-ac02-2f2f84f05b35", 00:07:53.101 "is_configured": true, 00:07:53.101 "data_offset": 0, 00:07:53.101 "data_size": 65536 00:07:53.101 } 00:07:53.101 ] 00:07:53.101 }' 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.101 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.360 [2024-12-07 05:35:26.692510] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:53.360 BaseBdev1 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.360 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.360 [ 00:07:53.360 { 00:07:53.360 "name": "BaseBdev1", 00:07:53.360 "aliases": [ 00:07:53.360 "f523d64f-bff5-4a5e-ae37-d9f8e65ceb79" 00:07:53.360 ], 00:07:53.360 "product_name": "Malloc disk", 00:07:53.360 "block_size": 512, 00:07:53.361 "num_blocks": 65536, 00:07:53.361 "uuid": "f523d64f-bff5-4a5e-ae37-d9f8e65ceb79", 00:07:53.361 "assigned_rate_limits": { 00:07:53.361 "rw_ios_per_sec": 0, 00:07:53.361 "rw_mbytes_per_sec": 0, 00:07:53.361 "r_mbytes_per_sec": 0, 00:07:53.361 "w_mbytes_per_sec": 0 00:07:53.361 }, 00:07:53.361 "claimed": true, 00:07:53.361 "claim_type": "exclusive_write", 00:07:53.361 "zoned": false, 00:07:53.361 "supported_io_types": { 00:07:53.361 "read": true, 00:07:53.361 "write": true, 00:07:53.361 "unmap": true, 00:07:53.361 "flush": true, 00:07:53.361 "reset": true, 00:07:53.361 "nvme_admin": false, 00:07:53.361 "nvme_io": false, 00:07:53.361 "nvme_io_md": false, 00:07:53.361 "write_zeroes": true, 00:07:53.361 "zcopy": true, 00:07:53.361 "get_zone_info": false, 00:07:53.361 "zone_management": false, 00:07:53.361 "zone_append": false, 00:07:53.361 "compare": false, 00:07:53.361 "compare_and_write": false, 00:07:53.361 "abort": true, 00:07:53.361 "seek_hole": false, 00:07:53.361 "seek_data": false, 00:07:53.361 "copy": true, 00:07:53.361 "nvme_iov_md": false 00:07:53.361 }, 00:07:53.621 "memory_domains": [ 00:07:53.621 { 00:07:53.621 "dma_device_id": "system", 00:07:53.621 "dma_device_type": 1 00:07:53.621 }, 00:07:53.621 { 00:07:53.621 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:53.621 "dma_device_type": 2 00:07:53.621 } 00:07:53.621 ], 00:07:53.621 "driver_specific": {} 00:07:53.621 } 00:07:53.621 ] 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.621 "name": "Existed_Raid", 00:07:53.621 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.621 "strip_size_kb": 64, 00:07:53.621 "state": "configuring", 00:07:53.621 "raid_level": "raid0", 00:07:53.621 "superblock": false, 00:07:53.621 "num_base_bdevs": 3, 00:07:53.621 "num_base_bdevs_discovered": 2, 00:07:53.621 "num_base_bdevs_operational": 3, 00:07:53.621 "base_bdevs_list": [ 00:07:53.621 { 00:07:53.621 "name": "BaseBdev1", 00:07:53.621 "uuid": "f523d64f-bff5-4a5e-ae37-d9f8e65ceb79", 00:07:53.621 "is_configured": true, 00:07:53.621 "data_offset": 0, 00:07:53.621 "data_size": 65536 00:07:53.621 }, 00:07:53.621 { 00:07:53.621 "name": null, 00:07:53.621 "uuid": "28903eab-c78e-4921-81b8-6458b54c5a6a", 00:07:53.621 "is_configured": false, 00:07:53.621 "data_offset": 0, 00:07:53.621 "data_size": 65536 00:07:53.621 }, 00:07:53.621 { 00:07:53.621 "name": "BaseBdev3", 00:07:53.621 "uuid": "0a91e3ae-7350-4acc-ac02-2f2f84f05b35", 00:07:53.621 "is_configured": true, 00:07:53.621 "data_offset": 0, 00:07:53.621 "data_size": 65536 00:07:53.621 } 00:07:53.621 ] 00:07:53.621 }' 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.621 05:35:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.881 [2024-12-07 05:35:27.083892] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.881 "name": "Existed_Raid", 00:07:53.881 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.881 "strip_size_kb": 64, 00:07:53.881 "state": "configuring", 00:07:53.881 "raid_level": "raid0", 00:07:53.881 "superblock": false, 00:07:53.881 "num_base_bdevs": 3, 00:07:53.881 "num_base_bdevs_discovered": 1, 00:07:53.881 "num_base_bdevs_operational": 3, 00:07:53.881 "base_bdevs_list": [ 00:07:53.881 { 00:07:53.881 "name": "BaseBdev1", 00:07:53.881 "uuid": "f523d64f-bff5-4a5e-ae37-d9f8e65ceb79", 00:07:53.881 "is_configured": true, 00:07:53.881 "data_offset": 0, 00:07:53.881 "data_size": 65536 00:07:53.881 }, 00:07:53.881 { 00:07:53.881 "name": null, 00:07:53.881 "uuid": "28903eab-c78e-4921-81b8-6458b54c5a6a", 00:07:53.881 "is_configured": false, 00:07:53.881 "data_offset": 0, 00:07:53.881 "data_size": 65536 00:07:53.881 }, 00:07:53.881 { 00:07:53.881 "name": null, 00:07:53.881 "uuid": "0a91e3ae-7350-4acc-ac02-2f2f84f05b35", 00:07:53.881 "is_configured": false, 00:07:53.881 "data_offset": 0, 00:07:53.881 "data_size": 65536 00:07:53.881 } 00:07:53.881 ] 00:07:53.881 }' 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.881 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.141 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:54.141 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.141 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.141 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.141 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.141 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:07:54.141 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:07:54.141 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.401 [2024-12-07 05:35:27.511179] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:54.401 "name": "Existed_Raid", 00:07:54.401 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.401 "strip_size_kb": 64, 00:07:54.401 "state": "configuring", 00:07:54.401 "raid_level": "raid0", 00:07:54.401 "superblock": false, 00:07:54.401 "num_base_bdevs": 3, 00:07:54.401 "num_base_bdevs_discovered": 2, 00:07:54.401 "num_base_bdevs_operational": 3, 00:07:54.401 "base_bdevs_list": [ 00:07:54.401 { 00:07:54.401 "name": "BaseBdev1", 00:07:54.401 "uuid": "f523d64f-bff5-4a5e-ae37-d9f8e65ceb79", 00:07:54.401 "is_configured": true, 00:07:54.401 "data_offset": 0, 00:07:54.401 "data_size": 65536 00:07:54.401 }, 00:07:54.401 { 00:07:54.401 "name": null, 00:07:54.401 "uuid": "28903eab-c78e-4921-81b8-6458b54c5a6a", 00:07:54.401 "is_configured": false, 00:07:54.401 "data_offset": 0, 00:07:54.401 "data_size": 65536 00:07:54.401 }, 00:07:54.401 { 00:07:54.401 "name": "BaseBdev3", 00:07:54.401 "uuid": "0a91e3ae-7350-4acc-ac02-2f2f84f05b35", 00:07:54.401 "is_configured": true, 00:07:54.401 "data_offset": 0, 00:07:54.401 "data_size": 65536 00:07:54.401 } 00:07:54.401 ] 00:07:54.401 }' 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:54.401 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.662 [2024-12-07 05:35:27.946541] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.662 05:35:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.662 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:54.662 "name": "Existed_Raid", 00:07:54.662 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.662 "strip_size_kb": 64, 00:07:54.662 "state": "configuring", 00:07:54.662 "raid_level": "raid0", 00:07:54.662 "superblock": false, 00:07:54.662 "num_base_bdevs": 3, 00:07:54.662 "num_base_bdevs_discovered": 1, 00:07:54.662 "num_base_bdevs_operational": 3, 00:07:54.662 "base_bdevs_list": [ 00:07:54.662 { 00:07:54.662 "name": null, 00:07:54.662 "uuid": "f523d64f-bff5-4a5e-ae37-d9f8e65ceb79", 00:07:54.662 "is_configured": false, 00:07:54.662 "data_offset": 0, 00:07:54.662 "data_size": 65536 00:07:54.662 }, 00:07:54.662 { 00:07:54.662 "name": null, 00:07:54.662 "uuid": "28903eab-c78e-4921-81b8-6458b54c5a6a", 00:07:54.662 "is_configured": false, 00:07:54.662 "data_offset": 0, 00:07:54.662 "data_size": 65536 00:07:54.662 }, 00:07:54.662 { 00:07:54.662 "name": "BaseBdev3", 00:07:54.662 "uuid": "0a91e3ae-7350-4acc-ac02-2f2f84f05b35", 00:07:54.662 "is_configured": true, 00:07:54.662 "data_offset": 0, 00:07:54.662 "data_size": 65536 00:07:54.662 } 00:07:54.662 ] 00:07:54.662 }' 00:07:54.662 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:54.662 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.232 [2024-12-07 05:35:28.420125] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.232 "name": "Existed_Raid", 00:07:55.232 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.232 "strip_size_kb": 64, 00:07:55.232 "state": "configuring", 00:07:55.232 "raid_level": "raid0", 00:07:55.232 "superblock": false, 00:07:55.232 "num_base_bdevs": 3, 00:07:55.232 "num_base_bdevs_discovered": 2, 00:07:55.232 "num_base_bdevs_operational": 3, 00:07:55.232 "base_bdevs_list": [ 00:07:55.232 { 00:07:55.232 "name": null, 00:07:55.232 "uuid": "f523d64f-bff5-4a5e-ae37-d9f8e65ceb79", 00:07:55.232 "is_configured": false, 00:07:55.232 "data_offset": 0, 00:07:55.232 "data_size": 65536 00:07:55.232 }, 00:07:55.232 { 00:07:55.232 "name": "BaseBdev2", 00:07:55.232 "uuid": "28903eab-c78e-4921-81b8-6458b54c5a6a", 00:07:55.232 "is_configured": true, 00:07:55.232 "data_offset": 0, 00:07:55.232 "data_size": 65536 00:07:55.232 }, 00:07:55.232 { 00:07:55.232 "name": "BaseBdev3", 00:07:55.232 "uuid": "0a91e3ae-7350-4acc-ac02-2f2f84f05b35", 00:07:55.232 "is_configured": true, 00:07:55.232 "data_offset": 0, 00:07:55.232 "data_size": 65536 00:07:55.232 } 00:07:55.232 ] 00:07:55.232 }' 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.232 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.492 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.492 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:07:55.492 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.492 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u f523d64f-bff5-4a5e-ae37-d9f8e65ceb79 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.754 [2024-12-07 05:35:28.938120] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:07:55.754 [2024-12-07 05:35:28.938245] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:07:55.754 [2024-12-07 05:35:28.938273] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:55.754 [2024-12-07 05:35:28.938546] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:55.754 [2024-12-07 05:35:28.938728] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:07:55.754 [2024-12-07 05:35:28.938770] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:07:55.754 [2024-12-07 05:35:28.939009] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:55.754 NewBaseBdev 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:07:55.754 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.755 [ 00:07:55.755 { 00:07:55.755 "name": "NewBaseBdev", 00:07:55.755 "aliases": [ 00:07:55.755 "f523d64f-bff5-4a5e-ae37-d9f8e65ceb79" 00:07:55.755 ], 00:07:55.755 "product_name": "Malloc disk", 00:07:55.755 "block_size": 512, 00:07:55.755 "num_blocks": 65536, 00:07:55.755 "uuid": "f523d64f-bff5-4a5e-ae37-d9f8e65ceb79", 00:07:55.755 "assigned_rate_limits": { 00:07:55.755 "rw_ios_per_sec": 0, 00:07:55.755 "rw_mbytes_per_sec": 0, 00:07:55.755 "r_mbytes_per_sec": 0, 00:07:55.755 "w_mbytes_per_sec": 0 00:07:55.755 }, 00:07:55.755 "claimed": true, 00:07:55.755 "claim_type": "exclusive_write", 00:07:55.755 "zoned": false, 00:07:55.755 "supported_io_types": { 00:07:55.755 "read": true, 00:07:55.755 "write": true, 00:07:55.755 "unmap": true, 00:07:55.755 "flush": true, 00:07:55.755 "reset": true, 00:07:55.755 "nvme_admin": false, 00:07:55.755 "nvme_io": false, 00:07:55.755 "nvme_io_md": false, 00:07:55.755 "write_zeroes": true, 00:07:55.755 "zcopy": true, 00:07:55.755 "get_zone_info": false, 00:07:55.755 "zone_management": false, 00:07:55.755 "zone_append": false, 00:07:55.755 "compare": false, 00:07:55.755 "compare_and_write": false, 00:07:55.755 "abort": true, 00:07:55.755 "seek_hole": false, 00:07:55.755 "seek_data": false, 00:07:55.755 "copy": true, 00:07:55.755 "nvme_iov_md": false 00:07:55.755 }, 00:07:55.755 "memory_domains": [ 00:07:55.755 { 00:07:55.755 "dma_device_id": "system", 00:07:55.755 "dma_device_type": 1 00:07:55.755 }, 00:07:55.755 { 00:07:55.755 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:55.755 "dma_device_type": 2 00:07:55.755 } 00:07:55.755 ], 00:07:55.755 "driver_specific": {} 00:07:55.755 } 00:07:55.755 ] 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.755 05:35:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.755 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.755 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.755 "name": "Existed_Raid", 00:07:55.755 "uuid": "3a893ee4-ba1a-46fc-a144-ba3c1d0cbbaa", 00:07:55.755 "strip_size_kb": 64, 00:07:55.755 "state": "online", 00:07:55.755 "raid_level": "raid0", 00:07:55.755 "superblock": false, 00:07:55.755 "num_base_bdevs": 3, 00:07:55.755 "num_base_bdevs_discovered": 3, 00:07:55.755 "num_base_bdevs_operational": 3, 00:07:55.755 "base_bdevs_list": [ 00:07:55.755 { 00:07:55.755 "name": "NewBaseBdev", 00:07:55.755 "uuid": "f523d64f-bff5-4a5e-ae37-d9f8e65ceb79", 00:07:55.755 "is_configured": true, 00:07:55.755 "data_offset": 0, 00:07:55.755 "data_size": 65536 00:07:55.755 }, 00:07:55.755 { 00:07:55.755 "name": "BaseBdev2", 00:07:55.755 "uuid": "28903eab-c78e-4921-81b8-6458b54c5a6a", 00:07:55.755 "is_configured": true, 00:07:55.755 "data_offset": 0, 00:07:55.755 "data_size": 65536 00:07:55.755 }, 00:07:55.755 { 00:07:55.755 "name": "BaseBdev3", 00:07:55.755 "uuid": "0a91e3ae-7350-4acc-ac02-2f2f84f05b35", 00:07:55.755 "is_configured": true, 00:07:55.755 "data_offset": 0, 00:07:55.755 "data_size": 65536 00:07:55.755 } 00:07:55.755 ] 00:07:55.755 }' 00:07:55.755 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.755 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.041 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:07:56.041 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:56.041 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:56.041 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:56.041 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:56.041 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:56.041 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:56.041 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:56.041 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.041 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.041 [2024-12-07 05:35:29.389745] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:56.312 "name": "Existed_Raid", 00:07:56.312 "aliases": [ 00:07:56.312 "3a893ee4-ba1a-46fc-a144-ba3c1d0cbbaa" 00:07:56.312 ], 00:07:56.312 "product_name": "Raid Volume", 00:07:56.312 "block_size": 512, 00:07:56.312 "num_blocks": 196608, 00:07:56.312 "uuid": "3a893ee4-ba1a-46fc-a144-ba3c1d0cbbaa", 00:07:56.312 "assigned_rate_limits": { 00:07:56.312 "rw_ios_per_sec": 0, 00:07:56.312 "rw_mbytes_per_sec": 0, 00:07:56.312 "r_mbytes_per_sec": 0, 00:07:56.312 "w_mbytes_per_sec": 0 00:07:56.312 }, 00:07:56.312 "claimed": false, 00:07:56.312 "zoned": false, 00:07:56.312 "supported_io_types": { 00:07:56.312 "read": true, 00:07:56.312 "write": true, 00:07:56.312 "unmap": true, 00:07:56.312 "flush": true, 00:07:56.312 "reset": true, 00:07:56.312 "nvme_admin": false, 00:07:56.312 "nvme_io": false, 00:07:56.312 "nvme_io_md": false, 00:07:56.312 "write_zeroes": true, 00:07:56.312 "zcopy": false, 00:07:56.312 "get_zone_info": false, 00:07:56.312 "zone_management": false, 00:07:56.312 "zone_append": false, 00:07:56.312 "compare": false, 00:07:56.312 "compare_and_write": false, 00:07:56.312 "abort": false, 00:07:56.312 "seek_hole": false, 00:07:56.312 "seek_data": false, 00:07:56.312 "copy": false, 00:07:56.312 "nvme_iov_md": false 00:07:56.312 }, 00:07:56.312 "memory_domains": [ 00:07:56.312 { 00:07:56.312 "dma_device_id": "system", 00:07:56.312 "dma_device_type": 1 00:07:56.312 }, 00:07:56.312 { 00:07:56.312 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.312 "dma_device_type": 2 00:07:56.312 }, 00:07:56.312 { 00:07:56.312 "dma_device_id": "system", 00:07:56.312 "dma_device_type": 1 00:07:56.312 }, 00:07:56.312 { 00:07:56.312 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.312 "dma_device_type": 2 00:07:56.312 }, 00:07:56.312 { 00:07:56.312 "dma_device_id": "system", 00:07:56.312 "dma_device_type": 1 00:07:56.312 }, 00:07:56.312 { 00:07:56.312 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.312 "dma_device_type": 2 00:07:56.312 } 00:07:56.312 ], 00:07:56.312 "driver_specific": { 00:07:56.312 "raid": { 00:07:56.312 "uuid": "3a893ee4-ba1a-46fc-a144-ba3c1d0cbbaa", 00:07:56.312 "strip_size_kb": 64, 00:07:56.312 "state": "online", 00:07:56.312 "raid_level": "raid0", 00:07:56.312 "superblock": false, 00:07:56.312 "num_base_bdevs": 3, 00:07:56.312 "num_base_bdevs_discovered": 3, 00:07:56.312 "num_base_bdevs_operational": 3, 00:07:56.312 "base_bdevs_list": [ 00:07:56.312 { 00:07:56.312 "name": "NewBaseBdev", 00:07:56.312 "uuid": "f523d64f-bff5-4a5e-ae37-d9f8e65ceb79", 00:07:56.312 "is_configured": true, 00:07:56.312 "data_offset": 0, 00:07:56.312 "data_size": 65536 00:07:56.312 }, 00:07:56.312 { 00:07:56.312 "name": "BaseBdev2", 00:07:56.312 "uuid": "28903eab-c78e-4921-81b8-6458b54c5a6a", 00:07:56.312 "is_configured": true, 00:07:56.312 "data_offset": 0, 00:07:56.312 "data_size": 65536 00:07:56.312 }, 00:07:56.312 { 00:07:56.312 "name": "BaseBdev3", 00:07:56.312 "uuid": "0a91e3ae-7350-4acc-ac02-2f2f84f05b35", 00:07:56.312 "is_configured": true, 00:07:56.312 "data_offset": 0, 00:07:56.312 "data_size": 65536 00:07:56.312 } 00:07:56.312 ] 00:07:56.312 } 00:07:56.312 } 00:07:56.312 }' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:07:56.312 BaseBdev2 00:07:56.312 BaseBdev3' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.312 [2024-12-07 05:35:29.660918] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:56.312 [2024-12-07 05:35:29.660987] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:56.312 [2024-12-07 05:35:29.661083] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:56.312 [2024-12-07 05:35:29.661162] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:56.312 [2024-12-07 05:35:29.661207] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 74718 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 74718 ']' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 74718 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:56.312 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74718 00:07:56.573 killing process with pid 74718 00:07:56.573 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:56.573 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:56.573 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74718' 00:07:56.573 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 74718 00:07:56.573 [2024-12-07 05:35:29.709190] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:56.573 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 74718 00:07:56.573 [2024-12-07 05:35:29.738762] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:56.833 05:35:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:56.833 ************************************ 00:07:56.833 END TEST raid_state_function_test 00:07:56.833 ************************************ 00:07:56.833 00:07:56.833 real 0m8.387s 00:07:56.833 user 0m14.355s 00:07:56.833 sys 0m1.580s 00:07:56.833 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:56.833 05:35:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.833 05:35:30 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:07:56.833 05:35:30 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:56.833 05:35:30 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:56.833 05:35:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:56.833 ************************************ 00:07:56.833 START TEST raid_state_function_test_sb 00:07:56.833 ************************************ 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 true 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=75317 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75317' 00:07:56.833 Process raid pid: 75317 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 75317 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 75317 ']' 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:56.833 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:56.833 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.833 [2024-12-07 05:35:30.110238] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:07:56.833 [2024-12-07 05:35:30.110464] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:57.092 [2024-12-07 05:35:30.264746] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:57.092 [2024-12-07 05:35:30.289275] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:57.092 [2024-12-07 05:35:30.330769] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:57.092 [2024-12-07 05:35:30.330807] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.659 [2024-12-07 05:35:30.940838] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:57.659 [2024-12-07 05:35:30.940941] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:57.659 [2024-12-07 05:35:30.940956] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:57.659 [2024-12-07 05:35:30.940966] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:57.659 [2024-12-07 05:35:30.940972] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:57.659 [2024-12-07 05:35:30.940984] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.659 "name": "Existed_Raid", 00:07:57.659 "uuid": "3e10ac2c-39c8-4fb4-b363-9356cd868ce8", 00:07:57.659 "strip_size_kb": 64, 00:07:57.659 "state": "configuring", 00:07:57.659 "raid_level": "raid0", 00:07:57.659 "superblock": true, 00:07:57.659 "num_base_bdevs": 3, 00:07:57.659 "num_base_bdevs_discovered": 0, 00:07:57.659 "num_base_bdevs_operational": 3, 00:07:57.659 "base_bdevs_list": [ 00:07:57.659 { 00:07:57.659 "name": "BaseBdev1", 00:07:57.659 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.659 "is_configured": false, 00:07:57.659 "data_offset": 0, 00:07:57.659 "data_size": 0 00:07:57.659 }, 00:07:57.659 { 00:07:57.659 "name": "BaseBdev2", 00:07:57.659 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.659 "is_configured": false, 00:07:57.659 "data_offset": 0, 00:07:57.659 "data_size": 0 00:07:57.659 }, 00:07:57.659 { 00:07:57.659 "name": "BaseBdev3", 00:07:57.659 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.659 "is_configured": false, 00:07:57.659 "data_offset": 0, 00:07:57.659 "data_size": 0 00:07:57.659 } 00:07:57.659 ] 00:07:57.659 }' 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.659 05:35:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.229 [2024-12-07 05:35:31.395913] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:58.229 [2024-12-07 05:35:31.395952] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.229 [2024-12-07 05:35:31.407915] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:58.229 [2024-12-07 05:35:31.407997] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:58.229 [2024-12-07 05:35:31.408027] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:58.229 [2024-12-07 05:35:31.408050] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:58.229 [2024-12-07 05:35:31.408078] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:58.229 [2024-12-07 05:35:31.408100] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.229 [2024-12-07 05:35:31.428663] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:58.229 BaseBdev1 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:58.229 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.230 [ 00:07:58.230 { 00:07:58.230 "name": "BaseBdev1", 00:07:58.230 "aliases": [ 00:07:58.230 "42db1300-6e4c-40bc-b6fe-37406094e2c7" 00:07:58.230 ], 00:07:58.230 "product_name": "Malloc disk", 00:07:58.230 "block_size": 512, 00:07:58.230 "num_blocks": 65536, 00:07:58.230 "uuid": "42db1300-6e4c-40bc-b6fe-37406094e2c7", 00:07:58.230 "assigned_rate_limits": { 00:07:58.230 "rw_ios_per_sec": 0, 00:07:58.230 "rw_mbytes_per_sec": 0, 00:07:58.230 "r_mbytes_per_sec": 0, 00:07:58.230 "w_mbytes_per_sec": 0 00:07:58.230 }, 00:07:58.230 "claimed": true, 00:07:58.230 "claim_type": "exclusive_write", 00:07:58.230 "zoned": false, 00:07:58.230 "supported_io_types": { 00:07:58.230 "read": true, 00:07:58.230 "write": true, 00:07:58.230 "unmap": true, 00:07:58.230 "flush": true, 00:07:58.230 "reset": true, 00:07:58.230 "nvme_admin": false, 00:07:58.230 "nvme_io": false, 00:07:58.230 "nvme_io_md": false, 00:07:58.230 "write_zeroes": true, 00:07:58.230 "zcopy": true, 00:07:58.230 "get_zone_info": false, 00:07:58.230 "zone_management": false, 00:07:58.230 "zone_append": false, 00:07:58.230 "compare": false, 00:07:58.230 "compare_and_write": false, 00:07:58.230 "abort": true, 00:07:58.230 "seek_hole": false, 00:07:58.230 "seek_data": false, 00:07:58.230 "copy": true, 00:07:58.230 "nvme_iov_md": false 00:07:58.230 }, 00:07:58.230 "memory_domains": [ 00:07:58.230 { 00:07:58.230 "dma_device_id": "system", 00:07:58.230 "dma_device_type": 1 00:07:58.230 }, 00:07:58.230 { 00:07:58.230 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:58.230 "dma_device_type": 2 00:07:58.230 } 00:07:58.230 ], 00:07:58.230 "driver_specific": {} 00:07:58.230 } 00:07:58.230 ] 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:58.230 "name": "Existed_Raid", 00:07:58.230 "uuid": "9da6c4e1-399f-4b8e-8be3-d1535b3eae1e", 00:07:58.230 "strip_size_kb": 64, 00:07:58.230 "state": "configuring", 00:07:58.230 "raid_level": "raid0", 00:07:58.230 "superblock": true, 00:07:58.230 "num_base_bdevs": 3, 00:07:58.230 "num_base_bdevs_discovered": 1, 00:07:58.230 "num_base_bdevs_operational": 3, 00:07:58.230 "base_bdevs_list": [ 00:07:58.230 { 00:07:58.230 "name": "BaseBdev1", 00:07:58.230 "uuid": "42db1300-6e4c-40bc-b6fe-37406094e2c7", 00:07:58.230 "is_configured": true, 00:07:58.230 "data_offset": 2048, 00:07:58.230 "data_size": 63488 00:07:58.230 }, 00:07:58.230 { 00:07:58.230 "name": "BaseBdev2", 00:07:58.230 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.230 "is_configured": false, 00:07:58.230 "data_offset": 0, 00:07:58.230 "data_size": 0 00:07:58.230 }, 00:07:58.230 { 00:07:58.230 "name": "BaseBdev3", 00:07:58.230 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.230 "is_configured": false, 00:07:58.230 "data_offset": 0, 00:07:58.230 "data_size": 0 00:07:58.230 } 00:07:58.230 ] 00:07:58.230 }' 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:58.230 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.801 [2024-12-07 05:35:31.939805] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:58.801 [2024-12-07 05:35:31.939912] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.801 [2024-12-07 05:35:31.951820] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:58.801 [2024-12-07 05:35:31.953635] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:58.801 [2024-12-07 05:35:31.953737] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:58.801 [2024-12-07 05:35:31.953751] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:58.801 [2024-12-07 05:35:31.953762] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.801 05:35:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.801 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:58.801 "name": "Existed_Raid", 00:07:58.801 "uuid": "e16718de-24f4-4b50-befc-91eaea2416b2", 00:07:58.801 "strip_size_kb": 64, 00:07:58.801 "state": "configuring", 00:07:58.801 "raid_level": "raid0", 00:07:58.801 "superblock": true, 00:07:58.801 "num_base_bdevs": 3, 00:07:58.801 "num_base_bdevs_discovered": 1, 00:07:58.801 "num_base_bdevs_operational": 3, 00:07:58.801 "base_bdevs_list": [ 00:07:58.801 { 00:07:58.801 "name": "BaseBdev1", 00:07:58.801 "uuid": "42db1300-6e4c-40bc-b6fe-37406094e2c7", 00:07:58.801 "is_configured": true, 00:07:58.801 "data_offset": 2048, 00:07:58.801 "data_size": 63488 00:07:58.801 }, 00:07:58.801 { 00:07:58.801 "name": "BaseBdev2", 00:07:58.801 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.801 "is_configured": false, 00:07:58.801 "data_offset": 0, 00:07:58.801 "data_size": 0 00:07:58.801 }, 00:07:58.801 { 00:07:58.801 "name": "BaseBdev3", 00:07:58.801 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.801 "is_configured": false, 00:07:58.801 "data_offset": 0, 00:07:58.801 "data_size": 0 00:07:58.801 } 00:07:58.801 ] 00:07:58.801 }' 00:07:58.801 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:58.801 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.061 [2024-12-07 05:35:32.413956] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:59.061 BaseBdev2 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.061 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.321 [ 00:07:59.321 { 00:07:59.321 "name": "BaseBdev2", 00:07:59.321 "aliases": [ 00:07:59.321 "5b8cfa2f-e94e-4ef3-8b84-6f4ad0b92dea" 00:07:59.321 ], 00:07:59.321 "product_name": "Malloc disk", 00:07:59.321 "block_size": 512, 00:07:59.321 "num_blocks": 65536, 00:07:59.321 "uuid": "5b8cfa2f-e94e-4ef3-8b84-6f4ad0b92dea", 00:07:59.321 "assigned_rate_limits": { 00:07:59.321 "rw_ios_per_sec": 0, 00:07:59.321 "rw_mbytes_per_sec": 0, 00:07:59.321 "r_mbytes_per_sec": 0, 00:07:59.321 "w_mbytes_per_sec": 0 00:07:59.321 }, 00:07:59.321 "claimed": true, 00:07:59.321 "claim_type": "exclusive_write", 00:07:59.321 "zoned": false, 00:07:59.321 "supported_io_types": { 00:07:59.321 "read": true, 00:07:59.321 "write": true, 00:07:59.321 "unmap": true, 00:07:59.321 "flush": true, 00:07:59.321 "reset": true, 00:07:59.321 "nvme_admin": false, 00:07:59.321 "nvme_io": false, 00:07:59.321 "nvme_io_md": false, 00:07:59.321 "write_zeroes": true, 00:07:59.321 "zcopy": true, 00:07:59.321 "get_zone_info": false, 00:07:59.321 "zone_management": false, 00:07:59.321 "zone_append": false, 00:07:59.321 "compare": false, 00:07:59.321 "compare_and_write": false, 00:07:59.321 "abort": true, 00:07:59.321 "seek_hole": false, 00:07:59.321 "seek_data": false, 00:07:59.321 "copy": true, 00:07:59.321 "nvme_iov_md": false 00:07:59.321 }, 00:07:59.321 "memory_domains": [ 00:07:59.321 { 00:07:59.321 "dma_device_id": "system", 00:07:59.321 "dma_device_type": 1 00:07:59.321 }, 00:07:59.321 { 00:07:59.321 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:59.321 "dma_device_type": 2 00:07:59.321 } 00:07:59.321 ], 00:07:59.321 "driver_specific": {} 00:07:59.321 } 00:07:59.321 ] 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.321 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.322 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.322 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.322 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.322 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:59.322 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.322 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.322 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.322 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.322 "name": "Existed_Raid", 00:07:59.322 "uuid": "e16718de-24f4-4b50-befc-91eaea2416b2", 00:07:59.322 "strip_size_kb": 64, 00:07:59.322 "state": "configuring", 00:07:59.322 "raid_level": "raid0", 00:07:59.322 "superblock": true, 00:07:59.322 "num_base_bdevs": 3, 00:07:59.322 "num_base_bdevs_discovered": 2, 00:07:59.322 "num_base_bdevs_operational": 3, 00:07:59.322 "base_bdevs_list": [ 00:07:59.322 { 00:07:59.322 "name": "BaseBdev1", 00:07:59.322 "uuid": "42db1300-6e4c-40bc-b6fe-37406094e2c7", 00:07:59.322 "is_configured": true, 00:07:59.322 "data_offset": 2048, 00:07:59.322 "data_size": 63488 00:07:59.322 }, 00:07:59.322 { 00:07:59.322 "name": "BaseBdev2", 00:07:59.322 "uuid": "5b8cfa2f-e94e-4ef3-8b84-6f4ad0b92dea", 00:07:59.322 "is_configured": true, 00:07:59.322 "data_offset": 2048, 00:07:59.322 "data_size": 63488 00:07:59.322 }, 00:07:59.322 { 00:07:59.322 "name": "BaseBdev3", 00:07:59.322 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:59.322 "is_configured": false, 00:07:59.322 "data_offset": 0, 00:07:59.322 "data_size": 0 00:07:59.322 } 00:07:59.322 ] 00:07:59.322 }' 00:07:59.322 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.322 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.581 [2024-12-07 05:35:32.867521] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:59.581 [2024-12-07 05:35:32.868130] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:59.581 [2024-12-07 05:35:32.868230] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:07:59.581 BaseBdev3 00:07:59.581 [2024-12-07 05:35:32.869177] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.581 [2024-12-07 05:35:32.869618] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:59.581 [2024-12-07 05:35:32.869694] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:07:59.581 [2024-12-07 05:35:32.870123] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.581 [ 00:07:59.581 { 00:07:59.581 "name": "BaseBdev3", 00:07:59.581 "aliases": [ 00:07:59.581 "da5a5554-896d-40d6-baed-3558572a3a24" 00:07:59.581 ], 00:07:59.581 "product_name": "Malloc disk", 00:07:59.581 "block_size": 512, 00:07:59.581 "num_blocks": 65536, 00:07:59.581 "uuid": "da5a5554-896d-40d6-baed-3558572a3a24", 00:07:59.581 "assigned_rate_limits": { 00:07:59.581 "rw_ios_per_sec": 0, 00:07:59.581 "rw_mbytes_per_sec": 0, 00:07:59.581 "r_mbytes_per_sec": 0, 00:07:59.581 "w_mbytes_per_sec": 0 00:07:59.581 }, 00:07:59.581 "claimed": true, 00:07:59.581 "claim_type": "exclusive_write", 00:07:59.581 "zoned": false, 00:07:59.581 "supported_io_types": { 00:07:59.581 "read": true, 00:07:59.581 "write": true, 00:07:59.581 "unmap": true, 00:07:59.581 "flush": true, 00:07:59.581 "reset": true, 00:07:59.581 "nvme_admin": false, 00:07:59.581 "nvme_io": false, 00:07:59.581 "nvme_io_md": false, 00:07:59.581 "write_zeroes": true, 00:07:59.581 "zcopy": true, 00:07:59.581 "get_zone_info": false, 00:07:59.581 "zone_management": false, 00:07:59.581 "zone_append": false, 00:07:59.581 "compare": false, 00:07:59.581 "compare_and_write": false, 00:07:59.581 "abort": true, 00:07:59.581 "seek_hole": false, 00:07:59.581 "seek_data": false, 00:07:59.581 "copy": true, 00:07:59.581 "nvme_iov_md": false 00:07:59.581 }, 00:07:59.581 "memory_domains": [ 00:07:59.581 { 00:07:59.581 "dma_device_id": "system", 00:07:59.581 "dma_device_type": 1 00:07:59.581 }, 00:07:59.581 { 00:07:59.581 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:59.581 "dma_device_type": 2 00:07:59.581 } 00:07:59.581 ], 00:07:59.581 "driver_specific": {} 00:07:59.581 } 00:07:59.581 ] 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.581 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.840 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.840 "name": "Existed_Raid", 00:07:59.840 "uuid": "e16718de-24f4-4b50-befc-91eaea2416b2", 00:07:59.840 "strip_size_kb": 64, 00:07:59.840 "state": "online", 00:07:59.840 "raid_level": "raid0", 00:07:59.840 "superblock": true, 00:07:59.840 "num_base_bdevs": 3, 00:07:59.840 "num_base_bdevs_discovered": 3, 00:07:59.840 "num_base_bdevs_operational": 3, 00:07:59.840 "base_bdevs_list": [ 00:07:59.840 { 00:07:59.840 "name": "BaseBdev1", 00:07:59.840 "uuid": "42db1300-6e4c-40bc-b6fe-37406094e2c7", 00:07:59.840 "is_configured": true, 00:07:59.840 "data_offset": 2048, 00:07:59.840 "data_size": 63488 00:07:59.840 }, 00:07:59.840 { 00:07:59.841 "name": "BaseBdev2", 00:07:59.841 "uuid": "5b8cfa2f-e94e-4ef3-8b84-6f4ad0b92dea", 00:07:59.841 "is_configured": true, 00:07:59.841 "data_offset": 2048, 00:07:59.841 "data_size": 63488 00:07:59.841 }, 00:07:59.841 { 00:07:59.841 "name": "BaseBdev3", 00:07:59.841 "uuid": "da5a5554-896d-40d6-baed-3558572a3a24", 00:07:59.841 "is_configured": true, 00:07:59.841 "data_offset": 2048, 00:07:59.841 "data_size": 63488 00:07:59.841 } 00:07:59.841 ] 00:07:59.841 }' 00:07:59.841 05:35:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.841 05:35:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.100 [2024-12-07 05:35:33.307025] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:00.100 "name": "Existed_Raid", 00:08:00.100 "aliases": [ 00:08:00.100 "e16718de-24f4-4b50-befc-91eaea2416b2" 00:08:00.100 ], 00:08:00.100 "product_name": "Raid Volume", 00:08:00.100 "block_size": 512, 00:08:00.100 "num_blocks": 190464, 00:08:00.100 "uuid": "e16718de-24f4-4b50-befc-91eaea2416b2", 00:08:00.100 "assigned_rate_limits": { 00:08:00.100 "rw_ios_per_sec": 0, 00:08:00.100 "rw_mbytes_per_sec": 0, 00:08:00.100 "r_mbytes_per_sec": 0, 00:08:00.100 "w_mbytes_per_sec": 0 00:08:00.100 }, 00:08:00.100 "claimed": false, 00:08:00.100 "zoned": false, 00:08:00.100 "supported_io_types": { 00:08:00.100 "read": true, 00:08:00.100 "write": true, 00:08:00.100 "unmap": true, 00:08:00.100 "flush": true, 00:08:00.100 "reset": true, 00:08:00.100 "nvme_admin": false, 00:08:00.100 "nvme_io": false, 00:08:00.100 "nvme_io_md": false, 00:08:00.100 "write_zeroes": true, 00:08:00.100 "zcopy": false, 00:08:00.100 "get_zone_info": false, 00:08:00.100 "zone_management": false, 00:08:00.100 "zone_append": false, 00:08:00.100 "compare": false, 00:08:00.100 "compare_and_write": false, 00:08:00.100 "abort": false, 00:08:00.100 "seek_hole": false, 00:08:00.100 "seek_data": false, 00:08:00.100 "copy": false, 00:08:00.100 "nvme_iov_md": false 00:08:00.100 }, 00:08:00.100 "memory_domains": [ 00:08:00.100 { 00:08:00.100 "dma_device_id": "system", 00:08:00.100 "dma_device_type": 1 00:08:00.100 }, 00:08:00.100 { 00:08:00.100 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:00.100 "dma_device_type": 2 00:08:00.100 }, 00:08:00.100 { 00:08:00.100 "dma_device_id": "system", 00:08:00.100 "dma_device_type": 1 00:08:00.100 }, 00:08:00.100 { 00:08:00.100 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:00.100 "dma_device_type": 2 00:08:00.100 }, 00:08:00.100 { 00:08:00.100 "dma_device_id": "system", 00:08:00.100 "dma_device_type": 1 00:08:00.100 }, 00:08:00.100 { 00:08:00.100 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:00.100 "dma_device_type": 2 00:08:00.100 } 00:08:00.100 ], 00:08:00.100 "driver_specific": { 00:08:00.100 "raid": { 00:08:00.100 "uuid": "e16718de-24f4-4b50-befc-91eaea2416b2", 00:08:00.100 "strip_size_kb": 64, 00:08:00.100 "state": "online", 00:08:00.100 "raid_level": "raid0", 00:08:00.100 "superblock": true, 00:08:00.100 "num_base_bdevs": 3, 00:08:00.100 "num_base_bdevs_discovered": 3, 00:08:00.100 "num_base_bdevs_operational": 3, 00:08:00.100 "base_bdevs_list": [ 00:08:00.100 { 00:08:00.100 "name": "BaseBdev1", 00:08:00.100 "uuid": "42db1300-6e4c-40bc-b6fe-37406094e2c7", 00:08:00.100 "is_configured": true, 00:08:00.100 "data_offset": 2048, 00:08:00.100 "data_size": 63488 00:08:00.100 }, 00:08:00.100 { 00:08:00.100 "name": "BaseBdev2", 00:08:00.100 "uuid": "5b8cfa2f-e94e-4ef3-8b84-6f4ad0b92dea", 00:08:00.100 "is_configured": true, 00:08:00.100 "data_offset": 2048, 00:08:00.100 "data_size": 63488 00:08:00.100 }, 00:08:00.100 { 00:08:00.100 "name": "BaseBdev3", 00:08:00.100 "uuid": "da5a5554-896d-40d6-baed-3558572a3a24", 00:08:00.100 "is_configured": true, 00:08:00.100 "data_offset": 2048, 00:08:00.100 "data_size": 63488 00:08:00.100 } 00:08:00.100 ] 00:08:00.100 } 00:08:00.100 } 00:08:00.100 }' 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:00.100 BaseBdev2 00:08:00.100 BaseBdev3' 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.100 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.359 [2024-12-07 05:35:33.582291] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:00.359 [2024-12-07 05:35:33.582361] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:00.359 [2024-12-07 05:35:33.582463] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:00.359 "name": "Existed_Raid", 00:08:00.359 "uuid": "e16718de-24f4-4b50-befc-91eaea2416b2", 00:08:00.359 "strip_size_kb": 64, 00:08:00.359 "state": "offline", 00:08:00.359 "raid_level": "raid0", 00:08:00.359 "superblock": true, 00:08:00.359 "num_base_bdevs": 3, 00:08:00.359 "num_base_bdevs_discovered": 2, 00:08:00.359 "num_base_bdevs_operational": 2, 00:08:00.359 "base_bdevs_list": [ 00:08:00.359 { 00:08:00.359 "name": null, 00:08:00.359 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:00.359 "is_configured": false, 00:08:00.359 "data_offset": 0, 00:08:00.359 "data_size": 63488 00:08:00.359 }, 00:08:00.359 { 00:08:00.359 "name": "BaseBdev2", 00:08:00.359 "uuid": "5b8cfa2f-e94e-4ef3-8b84-6f4ad0b92dea", 00:08:00.359 "is_configured": true, 00:08:00.359 "data_offset": 2048, 00:08:00.359 "data_size": 63488 00:08:00.359 }, 00:08:00.359 { 00:08:00.359 "name": "BaseBdev3", 00:08:00.359 "uuid": "da5a5554-896d-40d6-baed-3558572a3a24", 00:08:00.359 "is_configured": true, 00:08:00.359 "data_offset": 2048, 00:08:00.359 "data_size": 63488 00:08:00.359 } 00:08:00.359 ] 00:08:00.359 }' 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:00.359 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.619 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:00.619 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:00.619 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.619 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.619 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.619 05:35:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:00.878 05:35:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.878 [2024-12-07 05:35:34.032915] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.878 [2024-12-07 05:35:34.087862] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:00.878 [2024-12-07 05:35:34.087907] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.878 BaseBdev2 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:00.878 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.879 [ 00:08:00.879 { 00:08:00.879 "name": "BaseBdev2", 00:08:00.879 "aliases": [ 00:08:00.879 "12d5a076-de0d-4912-88fd-9ca69691cdf6" 00:08:00.879 ], 00:08:00.879 "product_name": "Malloc disk", 00:08:00.879 "block_size": 512, 00:08:00.879 "num_blocks": 65536, 00:08:00.879 "uuid": "12d5a076-de0d-4912-88fd-9ca69691cdf6", 00:08:00.879 "assigned_rate_limits": { 00:08:00.879 "rw_ios_per_sec": 0, 00:08:00.879 "rw_mbytes_per_sec": 0, 00:08:00.879 "r_mbytes_per_sec": 0, 00:08:00.879 "w_mbytes_per_sec": 0 00:08:00.879 }, 00:08:00.879 "claimed": false, 00:08:00.879 "zoned": false, 00:08:00.879 "supported_io_types": { 00:08:00.879 "read": true, 00:08:00.879 "write": true, 00:08:00.879 "unmap": true, 00:08:00.879 "flush": true, 00:08:00.879 "reset": true, 00:08:00.879 "nvme_admin": false, 00:08:00.879 "nvme_io": false, 00:08:00.879 "nvme_io_md": false, 00:08:00.879 "write_zeroes": true, 00:08:00.879 "zcopy": true, 00:08:00.879 "get_zone_info": false, 00:08:00.879 "zone_management": false, 00:08:00.879 "zone_append": false, 00:08:00.879 "compare": false, 00:08:00.879 "compare_and_write": false, 00:08:00.879 "abort": true, 00:08:00.879 "seek_hole": false, 00:08:00.879 "seek_data": false, 00:08:00.879 "copy": true, 00:08:00.879 "nvme_iov_md": false 00:08:00.879 }, 00:08:00.879 "memory_domains": [ 00:08:00.879 { 00:08:00.879 "dma_device_id": "system", 00:08:00.879 "dma_device_type": 1 00:08:00.879 }, 00:08:00.879 { 00:08:00.879 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:00.879 "dma_device_type": 2 00:08:00.879 } 00:08:00.879 ], 00:08:00.879 "driver_specific": {} 00:08:00.879 } 00:08:00.879 ] 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.879 BaseBdev3 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.879 [ 00:08:00.879 { 00:08:00.879 "name": "BaseBdev3", 00:08:00.879 "aliases": [ 00:08:00.879 "e7e2d91d-6a0c-4da4-a534-4e08980ad05d" 00:08:00.879 ], 00:08:00.879 "product_name": "Malloc disk", 00:08:00.879 "block_size": 512, 00:08:00.879 "num_blocks": 65536, 00:08:00.879 "uuid": "e7e2d91d-6a0c-4da4-a534-4e08980ad05d", 00:08:00.879 "assigned_rate_limits": { 00:08:00.879 "rw_ios_per_sec": 0, 00:08:00.879 "rw_mbytes_per_sec": 0, 00:08:00.879 "r_mbytes_per_sec": 0, 00:08:00.879 "w_mbytes_per_sec": 0 00:08:00.879 }, 00:08:00.879 "claimed": false, 00:08:00.879 "zoned": false, 00:08:00.879 "supported_io_types": { 00:08:00.879 "read": true, 00:08:00.879 "write": true, 00:08:00.879 "unmap": true, 00:08:00.879 "flush": true, 00:08:00.879 "reset": true, 00:08:00.879 "nvme_admin": false, 00:08:00.879 "nvme_io": false, 00:08:00.879 "nvme_io_md": false, 00:08:00.879 "write_zeroes": true, 00:08:00.879 "zcopy": true, 00:08:00.879 "get_zone_info": false, 00:08:00.879 "zone_management": false, 00:08:00.879 "zone_append": false, 00:08:00.879 "compare": false, 00:08:00.879 "compare_and_write": false, 00:08:00.879 "abort": true, 00:08:00.879 "seek_hole": false, 00:08:00.879 "seek_data": false, 00:08:00.879 "copy": true, 00:08:00.879 "nvme_iov_md": false 00:08:00.879 }, 00:08:00.879 "memory_domains": [ 00:08:00.879 { 00:08:00.879 "dma_device_id": "system", 00:08:00.879 "dma_device_type": 1 00:08:00.879 }, 00:08:00.879 { 00:08:00.879 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:00.879 "dma_device_type": 2 00:08:00.879 } 00:08:00.879 ], 00:08:00.879 "driver_specific": {} 00:08:00.879 } 00:08:00.879 ] 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.879 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.138 [2024-12-07 05:35:34.250098] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:01.138 [2024-12-07 05:35:34.250198] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:01.138 [2024-12-07 05:35:34.250242] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:01.138 [2024-12-07 05:35:34.252144] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:01.138 "name": "Existed_Raid", 00:08:01.138 "uuid": "29829d3f-752b-4d49-8d6f-9dcd7173ce7e", 00:08:01.138 "strip_size_kb": 64, 00:08:01.138 "state": "configuring", 00:08:01.138 "raid_level": "raid0", 00:08:01.138 "superblock": true, 00:08:01.138 "num_base_bdevs": 3, 00:08:01.138 "num_base_bdevs_discovered": 2, 00:08:01.138 "num_base_bdevs_operational": 3, 00:08:01.138 "base_bdevs_list": [ 00:08:01.138 { 00:08:01.138 "name": "BaseBdev1", 00:08:01.138 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:01.138 "is_configured": false, 00:08:01.138 "data_offset": 0, 00:08:01.138 "data_size": 0 00:08:01.138 }, 00:08:01.138 { 00:08:01.138 "name": "BaseBdev2", 00:08:01.138 "uuid": "12d5a076-de0d-4912-88fd-9ca69691cdf6", 00:08:01.138 "is_configured": true, 00:08:01.138 "data_offset": 2048, 00:08:01.138 "data_size": 63488 00:08:01.138 }, 00:08:01.138 { 00:08:01.138 "name": "BaseBdev3", 00:08:01.138 "uuid": "e7e2d91d-6a0c-4da4-a534-4e08980ad05d", 00:08:01.138 "is_configured": true, 00:08:01.138 "data_offset": 2048, 00:08:01.138 "data_size": 63488 00:08:01.138 } 00:08:01.138 ] 00:08:01.138 }' 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:01.138 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.396 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:01.396 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.396 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.396 [2024-12-07 05:35:34.709327] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:01.396 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.396 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:01.396 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:01.396 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:01.396 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:01.397 "name": "Existed_Raid", 00:08:01.397 "uuid": "29829d3f-752b-4d49-8d6f-9dcd7173ce7e", 00:08:01.397 "strip_size_kb": 64, 00:08:01.397 "state": "configuring", 00:08:01.397 "raid_level": "raid0", 00:08:01.397 "superblock": true, 00:08:01.397 "num_base_bdevs": 3, 00:08:01.397 "num_base_bdevs_discovered": 1, 00:08:01.397 "num_base_bdevs_operational": 3, 00:08:01.397 "base_bdevs_list": [ 00:08:01.397 { 00:08:01.397 "name": "BaseBdev1", 00:08:01.397 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:01.397 "is_configured": false, 00:08:01.397 "data_offset": 0, 00:08:01.397 "data_size": 0 00:08:01.397 }, 00:08:01.397 { 00:08:01.397 "name": null, 00:08:01.397 "uuid": "12d5a076-de0d-4912-88fd-9ca69691cdf6", 00:08:01.397 "is_configured": false, 00:08:01.397 "data_offset": 0, 00:08:01.397 "data_size": 63488 00:08:01.397 }, 00:08:01.397 { 00:08:01.397 "name": "BaseBdev3", 00:08:01.397 "uuid": "e7e2d91d-6a0c-4da4-a534-4e08980ad05d", 00:08:01.397 "is_configured": true, 00:08:01.397 "data_offset": 2048, 00:08:01.397 "data_size": 63488 00:08:01.397 } 00:08:01.397 ] 00:08:01.397 }' 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:01.397 05:35:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.964 [2024-12-07 05:35:35.167490] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:01.964 BaseBdev1 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.964 [ 00:08:01.964 { 00:08:01.964 "name": "BaseBdev1", 00:08:01.964 "aliases": [ 00:08:01.964 "9b7a4861-122d-4c03-a406-662b582cdb2d" 00:08:01.964 ], 00:08:01.964 "product_name": "Malloc disk", 00:08:01.964 "block_size": 512, 00:08:01.964 "num_blocks": 65536, 00:08:01.964 "uuid": "9b7a4861-122d-4c03-a406-662b582cdb2d", 00:08:01.964 "assigned_rate_limits": { 00:08:01.964 "rw_ios_per_sec": 0, 00:08:01.964 "rw_mbytes_per_sec": 0, 00:08:01.964 "r_mbytes_per_sec": 0, 00:08:01.964 "w_mbytes_per_sec": 0 00:08:01.964 }, 00:08:01.964 "claimed": true, 00:08:01.964 "claim_type": "exclusive_write", 00:08:01.964 "zoned": false, 00:08:01.964 "supported_io_types": { 00:08:01.964 "read": true, 00:08:01.964 "write": true, 00:08:01.964 "unmap": true, 00:08:01.964 "flush": true, 00:08:01.964 "reset": true, 00:08:01.964 "nvme_admin": false, 00:08:01.964 "nvme_io": false, 00:08:01.964 "nvme_io_md": false, 00:08:01.964 "write_zeroes": true, 00:08:01.964 "zcopy": true, 00:08:01.964 "get_zone_info": false, 00:08:01.964 "zone_management": false, 00:08:01.964 "zone_append": false, 00:08:01.964 "compare": false, 00:08:01.964 "compare_and_write": false, 00:08:01.964 "abort": true, 00:08:01.964 "seek_hole": false, 00:08:01.964 "seek_data": false, 00:08:01.964 "copy": true, 00:08:01.964 "nvme_iov_md": false 00:08:01.964 }, 00:08:01.964 "memory_domains": [ 00:08:01.964 { 00:08:01.964 "dma_device_id": "system", 00:08:01.964 "dma_device_type": 1 00:08:01.964 }, 00:08:01.964 { 00:08:01.964 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:01.964 "dma_device_type": 2 00:08:01.964 } 00:08:01.964 ], 00:08:01.964 "driver_specific": {} 00:08:01.964 } 00:08:01.964 ] 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.964 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:01.964 "name": "Existed_Raid", 00:08:01.964 "uuid": "29829d3f-752b-4d49-8d6f-9dcd7173ce7e", 00:08:01.964 "strip_size_kb": 64, 00:08:01.964 "state": "configuring", 00:08:01.964 "raid_level": "raid0", 00:08:01.964 "superblock": true, 00:08:01.964 "num_base_bdevs": 3, 00:08:01.964 "num_base_bdevs_discovered": 2, 00:08:01.964 "num_base_bdevs_operational": 3, 00:08:01.964 "base_bdevs_list": [ 00:08:01.964 { 00:08:01.964 "name": "BaseBdev1", 00:08:01.964 "uuid": "9b7a4861-122d-4c03-a406-662b582cdb2d", 00:08:01.964 "is_configured": true, 00:08:01.964 "data_offset": 2048, 00:08:01.964 "data_size": 63488 00:08:01.964 }, 00:08:01.964 { 00:08:01.964 "name": null, 00:08:01.964 "uuid": "12d5a076-de0d-4912-88fd-9ca69691cdf6", 00:08:01.964 "is_configured": false, 00:08:01.964 "data_offset": 0, 00:08:01.964 "data_size": 63488 00:08:01.964 }, 00:08:01.964 { 00:08:01.964 "name": "BaseBdev3", 00:08:01.964 "uuid": "e7e2d91d-6a0c-4da4-a534-4e08980ad05d", 00:08:01.964 "is_configured": true, 00:08:01.964 "data_offset": 2048, 00:08:01.964 "data_size": 63488 00:08:01.964 } 00:08:01.964 ] 00:08:01.964 }' 00:08:01.965 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:01.965 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.574 [2024-12-07 05:35:35.654730] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:02.574 "name": "Existed_Raid", 00:08:02.574 "uuid": "29829d3f-752b-4d49-8d6f-9dcd7173ce7e", 00:08:02.574 "strip_size_kb": 64, 00:08:02.574 "state": "configuring", 00:08:02.574 "raid_level": "raid0", 00:08:02.574 "superblock": true, 00:08:02.574 "num_base_bdevs": 3, 00:08:02.574 "num_base_bdevs_discovered": 1, 00:08:02.574 "num_base_bdevs_operational": 3, 00:08:02.574 "base_bdevs_list": [ 00:08:02.574 { 00:08:02.574 "name": "BaseBdev1", 00:08:02.574 "uuid": "9b7a4861-122d-4c03-a406-662b582cdb2d", 00:08:02.574 "is_configured": true, 00:08:02.574 "data_offset": 2048, 00:08:02.574 "data_size": 63488 00:08:02.574 }, 00:08:02.574 { 00:08:02.574 "name": null, 00:08:02.574 "uuid": "12d5a076-de0d-4912-88fd-9ca69691cdf6", 00:08:02.574 "is_configured": false, 00:08:02.574 "data_offset": 0, 00:08:02.574 "data_size": 63488 00:08:02.574 }, 00:08:02.574 { 00:08:02.574 "name": null, 00:08:02.574 "uuid": "e7e2d91d-6a0c-4da4-a534-4e08980ad05d", 00:08:02.574 "is_configured": false, 00:08:02.574 "data_offset": 0, 00:08:02.574 "data_size": 63488 00:08:02.574 } 00:08:02.574 ] 00:08:02.574 }' 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:02.574 05:35:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.832 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.832 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.832 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.832 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:02.832 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.832 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:02.832 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:02.832 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.832 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.833 [2024-12-07 05:35:36.185812] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.833 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.091 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.091 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:03.091 "name": "Existed_Raid", 00:08:03.091 "uuid": "29829d3f-752b-4d49-8d6f-9dcd7173ce7e", 00:08:03.091 "strip_size_kb": 64, 00:08:03.091 "state": "configuring", 00:08:03.091 "raid_level": "raid0", 00:08:03.091 "superblock": true, 00:08:03.091 "num_base_bdevs": 3, 00:08:03.091 "num_base_bdevs_discovered": 2, 00:08:03.091 "num_base_bdevs_operational": 3, 00:08:03.091 "base_bdevs_list": [ 00:08:03.091 { 00:08:03.091 "name": "BaseBdev1", 00:08:03.091 "uuid": "9b7a4861-122d-4c03-a406-662b582cdb2d", 00:08:03.091 "is_configured": true, 00:08:03.091 "data_offset": 2048, 00:08:03.091 "data_size": 63488 00:08:03.091 }, 00:08:03.091 { 00:08:03.091 "name": null, 00:08:03.091 "uuid": "12d5a076-de0d-4912-88fd-9ca69691cdf6", 00:08:03.091 "is_configured": false, 00:08:03.091 "data_offset": 0, 00:08:03.091 "data_size": 63488 00:08:03.091 }, 00:08:03.091 { 00:08:03.091 "name": "BaseBdev3", 00:08:03.091 "uuid": "e7e2d91d-6a0c-4da4-a534-4e08980ad05d", 00:08:03.091 "is_configured": true, 00:08:03.091 "data_offset": 2048, 00:08:03.091 "data_size": 63488 00:08:03.091 } 00:08:03.091 ] 00:08:03.091 }' 00:08:03.091 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:03.091 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.349 [2024-12-07 05:35:36.609120] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:03.349 "name": "Existed_Raid", 00:08:03.349 "uuid": "29829d3f-752b-4d49-8d6f-9dcd7173ce7e", 00:08:03.349 "strip_size_kb": 64, 00:08:03.349 "state": "configuring", 00:08:03.349 "raid_level": "raid0", 00:08:03.349 "superblock": true, 00:08:03.349 "num_base_bdevs": 3, 00:08:03.349 "num_base_bdevs_discovered": 1, 00:08:03.349 "num_base_bdevs_operational": 3, 00:08:03.349 "base_bdevs_list": [ 00:08:03.349 { 00:08:03.349 "name": null, 00:08:03.349 "uuid": "9b7a4861-122d-4c03-a406-662b582cdb2d", 00:08:03.349 "is_configured": false, 00:08:03.349 "data_offset": 0, 00:08:03.349 "data_size": 63488 00:08:03.349 }, 00:08:03.349 { 00:08:03.349 "name": null, 00:08:03.349 "uuid": "12d5a076-de0d-4912-88fd-9ca69691cdf6", 00:08:03.349 "is_configured": false, 00:08:03.349 "data_offset": 0, 00:08:03.349 "data_size": 63488 00:08:03.349 }, 00:08:03.349 { 00:08:03.349 "name": "BaseBdev3", 00:08:03.349 "uuid": "e7e2d91d-6a0c-4da4-a534-4e08980ad05d", 00:08:03.349 "is_configured": true, 00:08:03.349 "data_offset": 2048, 00:08:03.349 "data_size": 63488 00:08:03.349 } 00:08:03.349 ] 00:08:03.349 }' 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:03.349 05:35:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.916 [2024-12-07 05:35:37.122601] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:03.916 "name": "Existed_Raid", 00:08:03.916 "uuid": "29829d3f-752b-4d49-8d6f-9dcd7173ce7e", 00:08:03.916 "strip_size_kb": 64, 00:08:03.916 "state": "configuring", 00:08:03.916 "raid_level": "raid0", 00:08:03.916 "superblock": true, 00:08:03.916 "num_base_bdevs": 3, 00:08:03.916 "num_base_bdevs_discovered": 2, 00:08:03.916 "num_base_bdevs_operational": 3, 00:08:03.916 "base_bdevs_list": [ 00:08:03.916 { 00:08:03.916 "name": null, 00:08:03.916 "uuid": "9b7a4861-122d-4c03-a406-662b582cdb2d", 00:08:03.916 "is_configured": false, 00:08:03.916 "data_offset": 0, 00:08:03.916 "data_size": 63488 00:08:03.916 }, 00:08:03.916 { 00:08:03.916 "name": "BaseBdev2", 00:08:03.916 "uuid": "12d5a076-de0d-4912-88fd-9ca69691cdf6", 00:08:03.916 "is_configured": true, 00:08:03.916 "data_offset": 2048, 00:08:03.916 "data_size": 63488 00:08:03.916 }, 00:08:03.916 { 00:08:03.916 "name": "BaseBdev3", 00:08:03.916 "uuid": "e7e2d91d-6a0c-4da4-a534-4e08980ad05d", 00:08:03.916 "is_configured": true, 00:08:03.916 "data_offset": 2048, 00:08:03.916 "data_size": 63488 00:08:03.916 } 00:08:03.916 ] 00:08:03.916 }' 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:03.916 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 9b7a4861-122d-4c03-a406-662b582cdb2d 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.484 [2024-12-07 05:35:37.652566] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:04.484 [2024-12-07 05:35:37.652838] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:04.484 [2024-12-07 05:35:37.652891] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:04.484 [2024-12-07 05:35:37.653194] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:04.484 NewBaseBdev 00:08:04.484 [2024-12-07 05:35:37.653358] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:04.484 [2024-12-07 05:35:37.653372] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:04.484 [2024-12-07 05:35:37.653478] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.484 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.484 [ 00:08:04.484 { 00:08:04.484 "name": "NewBaseBdev", 00:08:04.484 "aliases": [ 00:08:04.484 "9b7a4861-122d-4c03-a406-662b582cdb2d" 00:08:04.484 ], 00:08:04.484 "product_name": "Malloc disk", 00:08:04.484 "block_size": 512, 00:08:04.484 "num_blocks": 65536, 00:08:04.484 "uuid": "9b7a4861-122d-4c03-a406-662b582cdb2d", 00:08:04.484 "assigned_rate_limits": { 00:08:04.484 "rw_ios_per_sec": 0, 00:08:04.484 "rw_mbytes_per_sec": 0, 00:08:04.484 "r_mbytes_per_sec": 0, 00:08:04.484 "w_mbytes_per_sec": 0 00:08:04.484 }, 00:08:04.484 "claimed": true, 00:08:04.484 "claim_type": "exclusive_write", 00:08:04.484 "zoned": false, 00:08:04.484 "supported_io_types": { 00:08:04.484 "read": true, 00:08:04.484 "write": true, 00:08:04.484 "unmap": true, 00:08:04.484 "flush": true, 00:08:04.484 "reset": true, 00:08:04.484 "nvme_admin": false, 00:08:04.484 "nvme_io": false, 00:08:04.484 "nvme_io_md": false, 00:08:04.484 "write_zeroes": true, 00:08:04.484 "zcopy": true, 00:08:04.484 "get_zone_info": false, 00:08:04.484 "zone_management": false, 00:08:04.484 "zone_append": false, 00:08:04.484 "compare": false, 00:08:04.484 "compare_and_write": false, 00:08:04.484 "abort": true, 00:08:04.484 "seek_hole": false, 00:08:04.484 "seek_data": false, 00:08:04.484 "copy": true, 00:08:04.484 "nvme_iov_md": false 00:08:04.484 }, 00:08:04.484 "memory_domains": [ 00:08:04.484 { 00:08:04.484 "dma_device_id": "system", 00:08:04.484 "dma_device_type": 1 00:08:04.484 }, 00:08:04.484 { 00:08:04.484 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:04.484 "dma_device_type": 2 00:08:04.484 } 00:08:04.484 ], 00:08:04.484 "driver_specific": {} 00:08:04.484 } 00:08:04.485 ] 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:04.485 "name": "Existed_Raid", 00:08:04.485 "uuid": "29829d3f-752b-4d49-8d6f-9dcd7173ce7e", 00:08:04.485 "strip_size_kb": 64, 00:08:04.485 "state": "online", 00:08:04.485 "raid_level": "raid0", 00:08:04.485 "superblock": true, 00:08:04.485 "num_base_bdevs": 3, 00:08:04.485 "num_base_bdevs_discovered": 3, 00:08:04.485 "num_base_bdevs_operational": 3, 00:08:04.485 "base_bdevs_list": [ 00:08:04.485 { 00:08:04.485 "name": "NewBaseBdev", 00:08:04.485 "uuid": "9b7a4861-122d-4c03-a406-662b582cdb2d", 00:08:04.485 "is_configured": true, 00:08:04.485 "data_offset": 2048, 00:08:04.485 "data_size": 63488 00:08:04.485 }, 00:08:04.485 { 00:08:04.485 "name": "BaseBdev2", 00:08:04.485 "uuid": "12d5a076-de0d-4912-88fd-9ca69691cdf6", 00:08:04.485 "is_configured": true, 00:08:04.485 "data_offset": 2048, 00:08:04.485 "data_size": 63488 00:08:04.485 }, 00:08:04.485 { 00:08:04.485 "name": "BaseBdev3", 00:08:04.485 "uuid": "e7e2d91d-6a0c-4da4-a534-4e08980ad05d", 00:08:04.485 "is_configured": true, 00:08:04.485 "data_offset": 2048, 00:08:04.485 "data_size": 63488 00:08:04.485 } 00:08:04.485 ] 00:08:04.485 }' 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:04.485 05:35:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.745 [2024-12-07 05:35:38.040208] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:04.745 "name": "Existed_Raid", 00:08:04.745 "aliases": [ 00:08:04.745 "29829d3f-752b-4d49-8d6f-9dcd7173ce7e" 00:08:04.745 ], 00:08:04.745 "product_name": "Raid Volume", 00:08:04.745 "block_size": 512, 00:08:04.745 "num_blocks": 190464, 00:08:04.745 "uuid": "29829d3f-752b-4d49-8d6f-9dcd7173ce7e", 00:08:04.745 "assigned_rate_limits": { 00:08:04.745 "rw_ios_per_sec": 0, 00:08:04.745 "rw_mbytes_per_sec": 0, 00:08:04.745 "r_mbytes_per_sec": 0, 00:08:04.745 "w_mbytes_per_sec": 0 00:08:04.745 }, 00:08:04.745 "claimed": false, 00:08:04.745 "zoned": false, 00:08:04.745 "supported_io_types": { 00:08:04.745 "read": true, 00:08:04.745 "write": true, 00:08:04.745 "unmap": true, 00:08:04.745 "flush": true, 00:08:04.745 "reset": true, 00:08:04.745 "nvme_admin": false, 00:08:04.745 "nvme_io": false, 00:08:04.745 "nvme_io_md": false, 00:08:04.745 "write_zeroes": true, 00:08:04.745 "zcopy": false, 00:08:04.745 "get_zone_info": false, 00:08:04.745 "zone_management": false, 00:08:04.745 "zone_append": false, 00:08:04.745 "compare": false, 00:08:04.745 "compare_and_write": false, 00:08:04.745 "abort": false, 00:08:04.745 "seek_hole": false, 00:08:04.745 "seek_data": false, 00:08:04.745 "copy": false, 00:08:04.745 "nvme_iov_md": false 00:08:04.745 }, 00:08:04.745 "memory_domains": [ 00:08:04.745 { 00:08:04.745 "dma_device_id": "system", 00:08:04.745 "dma_device_type": 1 00:08:04.745 }, 00:08:04.745 { 00:08:04.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:04.745 "dma_device_type": 2 00:08:04.745 }, 00:08:04.745 { 00:08:04.745 "dma_device_id": "system", 00:08:04.745 "dma_device_type": 1 00:08:04.745 }, 00:08:04.745 { 00:08:04.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:04.745 "dma_device_type": 2 00:08:04.745 }, 00:08:04.745 { 00:08:04.745 "dma_device_id": "system", 00:08:04.745 "dma_device_type": 1 00:08:04.745 }, 00:08:04.745 { 00:08:04.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:04.745 "dma_device_type": 2 00:08:04.745 } 00:08:04.745 ], 00:08:04.745 "driver_specific": { 00:08:04.745 "raid": { 00:08:04.745 "uuid": "29829d3f-752b-4d49-8d6f-9dcd7173ce7e", 00:08:04.745 "strip_size_kb": 64, 00:08:04.745 "state": "online", 00:08:04.745 "raid_level": "raid0", 00:08:04.745 "superblock": true, 00:08:04.745 "num_base_bdevs": 3, 00:08:04.745 "num_base_bdevs_discovered": 3, 00:08:04.745 "num_base_bdevs_operational": 3, 00:08:04.745 "base_bdevs_list": [ 00:08:04.745 { 00:08:04.745 "name": "NewBaseBdev", 00:08:04.745 "uuid": "9b7a4861-122d-4c03-a406-662b582cdb2d", 00:08:04.745 "is_configured": true, 00:08:04.745 "data_offset": 2048, 00:08:04.745 "data_size": 63488 00:08:04.745 }, 00:08:04.745 { 00:08:04.745 "name": "BaseBdev2", 00:08:04.745 "uuid": "12d5a076-de0d-4912-88fd-9ca69691cdf6", 00:08:04.745 "is_configured": true, 00:08:04.745 "data_offset": 2048, 00:08:04.745 "data_size": 63488 00:08:04.745 }, 00:08:04.745 { 00:08:04.745 "name": "BaseBdev3", 00:08:04.745 "uuid": "e7e2d91d-6a0c-4da4-a534-4e08980ad05d", 00:08:04.745 "is_configured": true, 00:08:04.745 "data_offset": 2048, 00:08:04.745 "data_size": 63488 00:08:04.745 } 00:08:04.745 ] 00:08:04.745 } 00:08:04.745 } 00:08:04.745 }' 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:04.745 BaseBdev2 00:08:04.745 BaseBdev3' 00:08:04.745 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.005 [2024-12-07 05:35:38.235577] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:05.005 [2024-12-07 05:35:38.235651] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:05.005 [2024-12-07 05:35:38.235739] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:05.005 [2024-12-07 05:35:38.235793] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:05.005 [2024-12-07 05:35:38.235805] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 75317 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 75317 ']' 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 75317 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75317 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75317' 00:08:05.005 killing process with pid 75317 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 75317 00:08:05.005 [2024-12-07 05:35:38.284893] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:05.005 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 75317 00:08:05.005 [2024-12-07 05:35:38.314541] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:05.264 05:35:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:05.264 00:08:05.264 real 0m8.494s 00:08:05.264 user 0m14.520s 00:08:05.264 sys 0m1.653s 00:08:05.264 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:05.264 ************************************ 00:08:05.264 END TEST raid_state_function_test_sb 00:08:05.264 ************************************ 00:08:05.264 05:35:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.264 05:35:38 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:08:05.264 05:35:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:08:05.264 05:35:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:05.264 05:35:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:05.264 ************************************ 00:08:05.264 START TEST raid_superblock_test 00:08:05.264 ************************************ 00:08:05.264 05:35:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 3 00:08:05.264 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:08:05.264 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:05.264 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:05.264 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:05.264 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:05.264 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=75921 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 75921 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 75921 ']' 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:05.265 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:05.265 05:35:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.524 [2024-12-07 05:35:38.666324] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:08:05.524 [2024-12-07 05:35:38.666549] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75921 ] 00:08:05.524 [2024-12-07 05:35:38.820487] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:05.524 [2024-12-07 05:35:38.845358] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:05.524 [2024-12-07 05:35:38.887375] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:05.524 [2024-12-07 05:35:38.887496] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.462 malloc1 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.462 [2024-12-07 05:35:39.514539] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:06.462 [2024-12-07 05:35:39.514669] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:06.462 [2024-12-07 05:35:39.514695] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:06.462 [2024-12-07 05:35:39.514715] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:06.462 [2024-12-07 05:35:39.516809] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:06.462 [2024-12-07 05:35:39.516846] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:06.462 pt1 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.462 malloc2 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:06.462 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.463 [2024-12-07 05:35:39.542941] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:06.463 [2024-12-07 05:35:39.543059] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:06.463 [2024-12-07 05:35:39.543096] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:06.463 [2024-12-07 05:35:39.543126] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:06.463 [2024-12-07 05:35:39.545198] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:06.463 [2024-12-07 05:35:39.545280] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:06.463 pt2 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.463 malloc3 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.463 [2024-12-07 05:35:39.575346] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:06.463 [2024-12-07 05:35:39.575458] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:06.463 [2024-12-07 05:35:39.575492] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:06.463 [2024-12-07 05:35:39.575522] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:06.463 [2024-12-07 05:35:39.577570] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:06.463 [2024-12-07 05:35:39.577649] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:06.463 pt3 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.463 [2024-12-07 05:35:39.587384] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:06.463 [2024-12-07 05:35:39.589242] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:06.463 [2024-12-07 05:35:39.589336] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:06.463 [2024-12-07 05:35:39.589509] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:06.463 [2024-12-07 05:35:39.589555] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:06.463 [2024-12-07 05:35:39.589825] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:06.463 [2024-12-07 05:35:39.589992] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:06.463 [2024-12-07 05:35:39.590038] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:06.463 [2024-12-07 05:35:39.590183] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:06.463 "name": "raid_bdev1", 00:08:06.463 "uuid": "a63783c4-0383-4b9b-9504-481573e0d1ae", 00:08:06.463 "strip_size_kb": 64, 00:08:06.463 "state": "online", 00:08:06.463 "raid_level": "raid0", 00:08:06.463 "superblock": true, 00:08:06.463 "num_base_bdevs": 3, 00:08:06.463 "num_base_bdevs_discovered": 3, 00:08:06.463 "num_base_bdevs_operational": 3, 00:08:06.463 "base_bdevs_list": [ 00:08:06.463 { 00:08:06.463 "name": "pt1", 00:08:06.463 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:06.463 "is_configured": true, 00:08:06.463 "data_offset": 2048, 00:08:06.463 "data_size": 63488 00:08:06.463 }, 00:08:06.463 { 00:08:06.463 "name": "pt2", 00:08:06.463 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:06.463 "is_configured": true, 00:08:06.463 "data_offset": 2048, 00:08:06.463 "data_size": 63488 00:08:06.463 }, 00:08:06.463 { 00:08:06.463 "name": "pt3", 00:08:06.463 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:06.463 "is_configured": true, 00:08:06.463 "data_offset": 2048, 00:08:06.463 "data_size": 63488 00:08:06.463 } 00:08:06.463 ] 00:08:06.463 }' 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:06.463 05:35:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.723 [2024-12-07 05:35:40.022897] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:06.723 "name": "raid_bdev1", 00:08:06.723 "aliases": [ 00:08:06.723 "a63783c4-0383-4b9b-9504-481573e0d1ae" 00:08:06.723 ], 00:08:06.723 "product_name": "Raid Volume", 00:08:06.723 "block_size": 512, 00:08:06.723 "num_blocks": 190464, 00:08:06.723 "uuid": "a63783c4-0383-4b9b-9504-481573e0d1ae", 00:08:06.723 "assigned_rate_limits": { 00:08:06.723 "rw_ios_per_sec": 0, 00:08:06.723 "rw_mbytes_per_sec": 0, 00:08:06.723 "r_mbytes_per_sec": 0, 00:08:06.723 "w_mbytes_per_sec": 0 00:08:06.723 }, 00:08:06.723 "claimed": false, 00:08:06.723 "zoned": false, 00:08:06.723 "supported_io_types": { 00:08:06.723 "read": true, 00:08:06.723 "write": true, 00:08:06.723 "unmap": true, 00:08:06.723 "flush": true, 00:08:06.723 "reset": true, 00:08:06.723 "nvme_admin": false, 00:08:06.723 "nvme_io": false, 00:08:06.723 "nvme_io_md": false, 00:08:06.723 "write_zeroes": true, 00:08:06.723 "zcopy": false, 00:08:06.723 "get_zone_info": false, 00:08:06.723 "zone_management": false, 00:08:06.723 "zone_append": false, 00:08:06.723 "compare": false, 00:08:06.723 "compare_and_write": false, 00:08:06.723 "abort": false, 00:08:06.723 "seek_hole": false, 00:08:06.723 "seek_data": false, 00:08:06.723 "copy": false, 00:08:06.723 "nvme_iov_md": false 00:08:06.723 }, 00:08:06.723 "memory_domains": [ 00:08:06.723 { 00:08:06.723 "dma_device_id": "system", 00:08:06.723 "dma_device_type": 1 00:08:06.723 }, 00:08:06.723 { 00:08:06.723 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:06.723 "dma_device_type": 2 00:08:06.723 }, 00:08:06.723 { 00:08:06.723 "dma_device_id": "system", 00:08:06.723 "dma_device_type": 1 00:08:06.723 }, 00:08:06.723 { 00:08:06.723 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:06.723 "dma_device_type": 2 00:08:06.723 }, 00:08:06.723 { 00:08:06.723 "dma_device_id": "system", 00:08:06.723 "dma_device_type": 1 00:08:06.723 }, 00:08:06.723 { 00:08:06.723 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:06.723 "dma_device_type": 2 00:08:06.723 } 00:08:06.723 ], 00:08:06.723 "driver_specific": { 00:08:06.723 "raid": { 00:08:06.723 "uuid": "a63783c4-0383-4b9b-9504-481573e0d1ae", 00:08:06.723 "strip_size_kb": 64, 00:08:06.723 "state": "online", 00:08:06.723 "raid_level": "raid0", 00:08:06.723 "superblock": true, 00:08:06.723 "num_base_bdevs": 3, 00:08:06.723 "num_base_bdevs_discovered": 3, 00:08:06.723 "num_base_bdevs_operational": 3, 00:08:06.723 "base_bdevs_list": [ 00:08:06.723 { 00:08:06.723 "name": "pt1", 00:08:06.723 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:06.723 "is_configured": true, 00:08:06.723 "data_offset": 2048, 00:08:06.723 "data_size": 63488 00:08:06.723 }, 00:08:06.723 { 00:08:06.723 "name": "pt2", 00:08:06.723 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:06.723 "is_configured": true, 00:08:06.723 "data_offset": 2048, 00:08:06.723 "data_size": 63488 00:08:06.723 }, 00:08:06.723 { 00:08:06.723 "name": "pt3", 00:08:06.723 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:06.723 "is_configured": true, 00:08:06.723 "data_offset": 2048, 00:08:06.723 "data_size": 63488 00:08:06.723 } 00:08:06.723 ] 00:08:06.723 } 00:08:06.723 } 00:08:06.723 }' 00:08:06.723 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:06.983 pt2 00:08:06.983 pt3' 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:06.983 [2024-12-07 05:35:40.302356] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=a63783c4-0383-4b9b-9504-481573e0d1ae 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z a63783c4-0383-4b9b-9504-481573e0d1ae ']' 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.983 [2024-12-07 05:35:40.342029] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:06.983 [2024-12-07 05:35:40.342060] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:06.983 [2024-12-07 05:35:40.342134] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:06.983 [2024-12-07 05:35:40.342190] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:06.983 [2024-12-07 05:35:40.342201] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:06.983 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.243 [2024-12-07 05:35:40.489819] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:07.243 [2024-12-07 05:35:40.491766] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:07.243 [2024-12-07 05:35:40.491853] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:07.243 [2024-12-07 05:35:40.491920] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:07.243 [2024-12-07 05:35:40.492003] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:07.243 [2024-12-07 05:35:40.492076] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:07.243 [2024-12-07 05:35:40.492114] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:07.243 [2024-12-07 05:35:40.492126] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:07.243 request: 00:08:07.243 { 00:08:07.243 "name": "raid_bdev1", 00:08:07.243 "raid_level": "raid0", 00:08:07.243 "base_bdevs": [ 00:08:07.243 "malloc1", 00:08:07.243 "malloc2", 00:08:07.243 "malloc3" 00:08:07.243 ], 00:08:07.243 "strip_size_kb": 64, 00:08:07.243 "superblock": false, 00:08:07.243 "method": "bdev_raid_create", 00:08:07.243 "req_id": 1 00:08:07.243 } 00:08:07.243 Got JSON-RPC error response 00:08:07.243 response: 00:08:07.243 { 00:08:07.243 "code": -17, 00:08:07.243 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:07.243 } 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.243 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.243 [2024-12-07 05:35:40.553674] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:07.243 [2024-12-07 05:35:40.553760] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:07.243 [2024-12-07 05:35:40.553790] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:07.244 [2024-12-07 05:35:40.553819] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:07.244 [2024-12-07 05:35:40.555966] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:07.244 [2024-12-07 05:35:40.556040] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:07.244 [2024-12-07 05:35:40.556123] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:07.244 [2024-12-07 05:35:40.556181] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:07.244 pt1 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:07.244 "name": "raid_bdev1", 00:08:07.244 "uuid": "a63783c4-0383-4b9b-9504-481573e0d1ae", 00:08:07.244 "strip_size_kb": 64, 00:08:07.244 "state": "configuring", 00:08:07.244 "raid_level": "raid0", 00:08:07.244 "superblock": true, 00:08:07.244 "num_base_bdevs": 3, 00:08:07.244 "num_base_bdevs_discovered": 1, 00:08:07.244 "num_base_bdevs_operational": 3, 00:08:07.244 "base_bdevs_list": [ 00:08:07.244 { 00:08:07.244 "name": "pt1", 00:08:07.244 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:07.244 "is_configured": true, 00:08:07.244 "data_offset": 2048, 00:08:07.244 "data_size": 63488 00:08:07.244 }, 00:08:07.244 { 00:08:07.244 "name": null, 00:08:07.244 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:07.244 "is_configured": false, 00:08:07.244 "data_offset": 2048, 00:08:07.244 "data_size": 63488 00:08:07.244 }, 00:08:07.244 { 00:08:07.244 "name": null, 00:08:07.244 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:07.244 "is_configured": false, 00:08:07.244 "data_offset": 2048, 00:08:07.244 "data_size": 63488 00:08:07.244 } 00:08:07.244 ] 00:08:07.244 }' 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:07.244 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.815 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:07.815 05:35:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:07.815 05:35:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.815 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.815 [2024-12-07 05:35:41.004895] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:07.815 [2024-12-07 05:35:41.004958] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:07.815 [2024-12-07 05:35:41.004977] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:07.815 [2024-12-07 05:35:41.004989] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:07.815 [2024-12-07 05:35:41.005338] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:07.815 [2024-12-07 05:35:41.005357] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:07.815 [2024-12-07 05:35:41.005417] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:07.815 [2024-12-07 05:35:41.005438] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:07.815 pt2 00:08:07.815 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.815 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:07.815 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.815 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.815 [2024-12-07 05:35:41.016888] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:07.815 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.815 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:07.816 "name": "raid_bdev1", 00:08:07.816 "uuid": "a63783c4-0383-4b9b-9504-481573e0d1ae", 00:08:07.816 "strip_size_kb": 64, 00:08:07.816 "state": "configuring", 00:08:07.816 "raid_level": "raid0", 00:08:07.816 "superblock": true, 00:08:07.816 "num_base_bdevs": 3, 00:08:07.816 "num_base_bdevs_discovered": 1, 00:08:07.816 "num_base_bdevs_operational": 3, 00:08:07.816 "base_bdevs_list": [ 00:08:07.816 { 00:08:07.816 "name": "pt1", 00:08:07.816 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:07.816 "is_configured": true, 00:08:07.816 "data_offset": 2048, 00:08:07.816 "data_size": 63488 00:08:07.816 }, 00:08:07.816 { 00:08:07.816 "name": null, 00:08:07.816 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:07.816 "is_configured": false, 00:08:07.816 "data_offset": 0, 00:08:07.816 "data_size": 63488 00:08:07.816 }, 00:08:07.816 { 00:08:07.816 "name": null, 00:08:07.816 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:07.816 "is_configured": false, 00:08:07.816 "data_offset": 2048, 00:08:07.816 "data_size": 63488 00:08:07.816 } 00:08:07.816 ] 00:08:07.816 }' 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:07.816 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.076 [2024-12-07 05:35:41.392239] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:08.076 [2024-12-07 05:35:41.392335] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:08.076 [2024-12-07 05:35:41.392370] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:08.076 [2024-12-07 05:35:41.392396] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:08.076 [2024-12-07 05:35:41.392785] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:08.076 [2024-12-07 05:35:41.392837] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:08.076 [2024-12-07 05:35:41.392927] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:08.076 [2024-12-07 05:35:41.392973] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:08.076 pt2 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.076 [2024-12-07 05:35:41.400224] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:08.076 [2024-12-07 05:35:41.400318] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:08.076 [2024-12-07 05:35:41.400351] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:08.076 [2024-12-07 05:35:41.400377] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:08.076 [2024-12-07 05:35:41.400712] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:08.076 [2024-12-07 05:35:41.400762] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:08.076 [2024-12-07 05:35:41.400838] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:08.076 [2024-12-07 05:35:41.400886] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:08.076 [2024-12-07 05:35:41.401004] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:08.076 [2024-12-07 05:35:41.401044] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:08.076 [2024-12-07 05:35:41.401277] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:08.076 [2024-12-07 05:35:41.401419] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:08.076 [2024-12-07 05:35:41.401462] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:08.076 [2024-12-07 05:35:41.401596] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:08.076 pt3 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:08.076 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.077 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.077 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.077 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.077 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:08.077 "name": "raid_bdev1", 00:08:08.077 "uuid": "a63783c4-0383-4b9b-9504-481573e0d1ae", 00:08:08.077 "strip_size_kb": 64, 00:08:08.077 "state": "online", 00:08:08.077 "raid_level": "raid0", 00:08:08.077 "superblock": true, 00:08:08.077 "num_base_bdevs": 3, 00:08:08.077 "num_base_bdevs_discovered": 3, 00:08:08.077 "num_base_bdevs_operational": 3, 00:08:08.077 "base_bdevs_list": [ 00:08:08.077 { 00:08:08.077 "name": "pt1", 00:08:08.077 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:08.077 "is_configured": true, 00:08:08.077 "data_offset": 2048, 00:08:08.077 "data_size": 63488 00:08:08.077 }, 00:08:08.077 { 00:08:08.077 "name": "pt2", 00:08:08.077 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:08.077 "is_configured": true, 00:08:08.077 "data_offset": 2048, 00:08:08.077 "data_size": 63488 00:08:08.077 }, 00:08:08.077 { 00:08:08.077 "name": "pt3", 00:08:08.077 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:08.077 "is_configured": true, 00:08:08.077 "data_offset": 2048, 00:08:08.077 "data_size": 63488 00:08:08.077 } 00:08:08.077 ] 00:08:08.077 }' 00:08:08.077 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:08.077 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.646 [2024-12-07 05:35:41.799854] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.646 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:08.646 "name": "raid_bdev1", 00:08:08.646 "aliases": [ 00:08:08.646 "a63783c4-0383-4b9b-9504-481573e0d1ae" 00:08:08.646 ], 00:08:08.646 "product_name": "Raid Volume", 00:08:08.646 "block_size": 512, 00:08:08.646 "num_blocks": 190464, 00:08:08.646 "uuid": "a63783c4-0383-4b9b-9504-481573e0d1ae", 00:08:08.646 "assigned_rate_limits": { 00:08:08.646 "rw_ios_per_sec": 0, 00:08:08.646 "rw_mbytes_per_sec": 0, 00:08:08.646 "r_mbytes_per_sec": 0, 00:08:08.646 "w_mbytes_per_sec": 0 00:08:08.646 }, 00:08:08.646 "claimed": false, 00:08:08.646 "zoned": false, 00:08:08.646 "supported_io_types": { 00:08:08.646 "read": true, 00:08:08.646 "write": true, 00:08:08.646 "unmap": true, 00:08:08.646 "flush": true, 00:08:08.646 "reset": true, 00:08:08.646 "nvme_admin": false, 00:08:08.646 "nvme_io": false, 00:08:08.646 "nvme_io_md": false, 00:08:08.646 "write_zeroes": true, 00:08:08.646 "zcopy": false, 00:08:08.646 "get_zone_info": false, 00:08:08.646 "zone_management": false, 00:08:08.647 "zone_append": false, 00:08:08.647 "compare": false, 00:08:08.647 "compare_and_write": false, 00:08:08.647 "abort": false, 00:08:08.647 "seek_hole": false, 00:08:08.647 "seek_data": false, 00:08:08.647 "copy": false, 00:08:08.647 "nvme_iov_md": false 00:08:08.647 }, 00:08:08.647 "memory_domains": [ 00:08:08.647 { 00:08:08.647 "dma_device_id": "system", 00:08:08.647 "dma_device_type": 1 00:08:08.647 }, 00:08:08.647 { 00:08:08.647 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:08.647 "dma_device_type": 2 00:08:08.647 }, 00:08:08.647 { 00:08:08.647 "dma_device_id": "system", 00:08:08.647 "dma_device_type": 1 00:08:08.647 }, 00:08:08.647 { 00:08:08.647 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:08.647 "dma_device_type": 2 00:08:08.647 }, 00:08:08.647 { 00:08:08.647 "dma_device_id": "system", 00:08:08.647 "dma_device_type": 1 00:08:08.647 }, 00:08:08.647 { 00:08:08.647 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:08.647 "dma_device_type": 2 00:08:08.647 } 00:08:08.647 ], 00:08:08.647 "driver_specific": { 00:08:08.647 "raid": { 00:08:08.647 "uuid": "a63783c4-0383-4b9b-9504-481573e0d1ae", 00:08:08.647 "strip_size_kb": 64, 00:08:08.647 "state": "online", 00:08:08.647 "raid_level": "raid0", 00:08:08.647 "superblock": true, 00:08:08.647 "num_base_bdevs": 3, 00:08:08.647 "num_base_bdevs_discovered": 3, 00:08:08.647 "num_base_bdevs_operational": 3, 00:08:08.647 "base_bdevs_list": [ 00:08:08.647 { 00:08:08.647 "name": "pt1", 00:08:08.647 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:08.647 "is_configured": true, 00:08:08.647 "data_offset": 2048, 00:08:08.647 "data_size": 63488 00:08:08.647 }, 00:08:08.647 { 00:08:08.647 "name": "pt2", 00:08:08.647 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:08.647 "is_configured": true, 00:08:08.647 "data_offset": 2048, 00:08:08.647 "data_size": 63488 00:08:08.647 }, 00:08:08.647 { 00:08:08.647 "name": "pt3", 00:08:08.647 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:08.647 "is_configured": true, 00:08:08.647 "data_offset": 2048, 00:08:08.647 "data_size": 63488 00:08:08.647 } 00:08:08.647 ] 00:08:08.647 } 00:08:08.647 } 00:08:08.647 }' 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:08.647 pt2 00:08:08.647 pt3' 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.647 05:35:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.906 [2024-12-07 05:35:42.047347] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' a63783c4-0383-4b9b-9504-481573e0d1ae '!=' a63783c4-0383-4b9b-9504-481573e0d1ae ']' 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 75921 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 75921 ']' 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 75921 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75921 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75921' 00:08:08.906 killing process with pid 75921 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 75921 00:08:08.906 [2024-12-07 05:35:42.121206] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:08.906 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 75921 00:08:08.906 [2024-12-07 05:35:42.121330] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:08.906 [2024-12-07 05:35:42.121396] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:08.906 [2024-12-07 05:35:42.121409] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:08.906 [2024-12-07 05:35:42.153214] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:09.165 05:35:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:09.165 00:08:09.165 real 0m3.778s 00:08:09.165 user 0m5.982s 00:08:09.165 sys 0m0.767s 00:08:09.165 ************************************ 00:08:09.165 END TEST raid_superblock_test 00:08:09.165 ************************************ 00:08:09.165 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:09.165 05:35:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.165 05:35:42 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:08:09.165 05:35:42 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:09.165 05:35:42 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:09.165 05:35:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:09.165 ************************************ 00:08:09.165 START TEST raid_read_error_test 00:08:09.165 ************************************ 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 read 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.1d4RIH6Qd2 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76162 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76162 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 76162 ']' 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:09.165 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:09.165 05:35:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.424 [2024-12-07 05:35:42.539853] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:08:09.424 [2024-12-07 05:35:42.540001] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76162 ] 00:08:09.424 [2024-12-07 05:35:42.696066] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:09.424 [2024-12-07 05:35:42.720786] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:09.424 [2024-12-07 05:35:42.762064] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:09.424 [2024-12-07 05:35:42.762184] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:09.992 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:09.992 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:09.992 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:09.992 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:09.992 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.992 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.252 BaseBdev1_malloc 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.252 true 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.252 [2024-12-07 05:35:43.388769] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:10.252 [2024-12-07 05:35:43.388823] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:10.252 [2024-12-07 05:35:43.388846] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:10.252 [2024-12-07 05:35:43.388856] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:10.252 [2024-12-07 05:35:43.390947] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:10.252 [2024-12-07 05:35:43.391061] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:10.252 BaseBdev1 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.252 BaseBdev2_malloc 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.252 true 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.252 [2024-12-07 05:35:43.429207] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:10.252 [2024-12-07 05:35:43.429299] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:10.252 [2024-12-07 05:35:43.429320] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:10.252 [2024-12-07 05:35:43.429337] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:10.252 [2024-12-07 05:35:43.431409] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:10.252 [2024-12-07 05:35:43.431447] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:10.252 BaseBdev2 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.252 BaseBdev3_malloc 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.252 true 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.252 [2024-12-07 05:35:43.469512] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:10.252 [2024-12-07 05:35:43.469559] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:10.252 [2024-12-07 05:35:43.469591] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:10.252 [2024-12-07 05:35:43.469599] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:10.252 [2024-12-07 05:35:43.471670] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:10.252 [2024-12-07 05:35:43.471704] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:10.252 BaseBdev3 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.252 [2024-12-07 05:35:43.481546] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:10.252 [2024-12-07 05:35:43.483357] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:10.252 [2024-12-07 05:35:43.483485] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:10.252 [2024-12-07 05:35:43.483677] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:10.252 [2024-12-07 05:35:43.483693] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:10.252 [2024-12-07 05:35:43.483949] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:08:10.252 [2024-12-07 05:35:43.484082] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:10.252 [2024-12-07 05:35:43.484098] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:10.252 [2024-12-07 05:35:43.484221] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:10.252 "name": "raid_bdev1", 00:08:10.252 "uuid": "7442bd00-b363-4292-9998-55d051c3b2f6", 00:08:10.252 "strip_size_kb": 64, 00:08:10.252 "state": "online", 00:08:10.252 "raid_level": "raid0", 00:08:10.252 "superblock": true, 00:08:10.252 "num_base_bdevs": 3, 00:08:10.252 "num_base_bdevs_discovered": 3, 00:08:10.252 "num_base_bdevs_operational": 3, 00:08:10.252 "base_bdevs_list": [ 00:08:10.252 { 00:08:10.252 "name": "BaseBdev1", 00:08:10.252 "uuid": "9878d353-7657-5c86-88d1-2d214acdcd52", 00:08:10.252 "is_configured": true, 00:08:10.252 "data_offset": 2048, 00:08:10.252 "data_size": 63488 00:08:10.252 }, 00:08:10.252 { 00:08:10.252 "name": "BaseBdev2", 00:08:10.252 "uuid": "f5ecc256-7c80-5ab2-8883-f715f65d2bf7", 00:08:10.252 "is_configured": true, 00:08:10.252 "data_offset": 2048, 00:08:10.252 "data_size": 63488 00:08:10.252 }, 00:08:10.252 { 00:08:10.252 "name": "BaseBdev3", 00:08:10.252 "uuid": "28a13243-25fe-5201-858e-8822a9799a63", 00:08:10.252 "is_configured": true, 00:08:10.252 "data_offset": 2048, 00:08:10.252 "data_size": 63488 00:08:10.252 } 00:08:10.252 ] 00:08:10.252 }' 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:10.252 05:35:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.820 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:10.820 05:35:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:10.820 [2024-12-07 05:35:44.016998] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:11.758 "name": "raid_bdev1", 00:08:11.758 "uuid": "7442bd00-b363-4292-9998-55d051c3b2f6", 00:08:11.758 "strip_size_kb": 64, 00:08:11.758 "state": "online", 00:08:11.758 "raid_level": "raid0", 00:08:11.758 "superblock": true, 00:08:11.758 "num_base_bdevs": 3, 00:08:11.758 "num_base_bdevs_discovered": 3, 00:08:11.758 "num_base_bdevs_operational": 3, 00:08:11.758 "base_bdevs_list": [ 00:08:11.758 { 00:08:11.758 "name": "BaseBdev1", 00:08:11.758 "uuid": "9878d353-7657-5c86-88d1-2d214acdcd52", 00:08:11.758 "is_configured": true, 00:08:11.758 "data_offset": 2048, 00:08:11.758 "data_size": 63488 00:08:11.758 }, 00:08:11.758 { 00:08:11.758 "name": "BaseBdev2", 00:08:11.758 "uuid": "f5ecc256-7c80-5ab2-8883-f715f65d2bf7", 00:08:11.758 "is_configured": true, 00:08:11.758 "data_offset": 2048, 00:08:11.758 "data_size": 63488 00:08:11.758 }, 00:08:11.758 { 00:08:11.758 "name": "BaseBdev3", 00:08:11.758 "uuid": "28a13243-25fe-5201-858e-8822a9799a63", 00:08:11.758 "is_configured": true, 00:08:11.758 "data_offset": 2048, 00:08:11.758 "data_size": 63488 00:08:11.758 } 00:08:11.758 ] 00:08:11.758 }' 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:11.758 05:35:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.016 05:35:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:12.016 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.016 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.016 [2024-12-07 05:35:45.364824] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:12.016 [2024-12-07 05:35:45.364915] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:12.016 [2024-12-07 05:35:45.367543] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:12.016 [2024-12-07 05:35:45.367643] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:12.016 [2024-12-07 05:35:45.367699] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:12.016 [2024-12-07 05:35:45.367755] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:12.016 { 00:08:12.016 "results": [ 00:08:12.016 { 00:08:12.016 "job": "raid_bdev1", 00:08:12.016 "core_mask": "0x1", 00:08:12.016 "workload": "randrw", 00:08:12.016 "percentage": 50, 00:08:12.016 "status": "finished", 00:08:12.016 "queue_depth": 1, 00:08:12.016 "io_size": 131072, 00:08:12.016 "runtime": 1.348778, 00:08:12.016 "iops": 16755.166528516926, 00:08:12.016 "mibps": 2094.3958160646157, 00:08:12.016 "io_failed": 1, 00:08:12.016 "io_timeout": 0, 00:08:12.016 "avg_latency_us": 82.51342644046836, 00:08:12.016 "min_latency_us": 20.90480349344978, 00:08:12.016 "max_latency_us": 1373.6803493449781 00:08:12.016 } 00:08:12.016 ], 00:08:12.016 "core_count": 1 00:08:12.016 } 00:08:12.016 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.016 05:35:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76162 00:08:12.016 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 76162 ']' 00:08:12.016 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 76162 00:08:12.016 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:08:12.016 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:12.016 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76162 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76162' 00:08:12.274 killing process with pid 76162 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 76162 00:08:12.274 [2024-12-07 05:35:45.400393] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 76162 00:08:12.274 [2024-12-07 05:35:45.424584] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.1d4RIH6Qd2 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:12.274 ************************************ 00:08:12.274 END TEST raid_read_error_test 00:08:12.274 ************************************ 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:08:12.274 00:08:12.274 real 0m3.200s 00:08:12.274 user 0m4.062s 00:08:12.274 sys 0m0.489s 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:12.274 05:35:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.532 05:35:45 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:08:12.532 05:35:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:12.532 05:35:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:12.532 05:35:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:12.532 ************************************ 00:08:12.532 START TEST raid_write_error_test 00:08:12.532 ************************************ 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 write 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.JqNlcoDMXK 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76292 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76292 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 76292 ']' 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:12.532 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:12.532 05:35:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.532 [2024-12-07 05:35:45.807496] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:08:12.532 [2024-12-07 05:35:45.807608] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76292 ] 00:08:12.790 [2024-12-07 05:35:45.961577] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:12.790 [2024-12-07 05:35:45.986180] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:12.790 [2024-12-07 05:35:46.028018] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:12.790 [2024-12-07 05:35:46.028053] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.358 BaseBdev1_malloc 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.358 true 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.358 [2024-12-07 05:35:46.659143] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:13.358 [2024-12-07 05:35:46.659202] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:13.358 [2024-12-07 05:35:46.659222] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:13.358 [2024-12-07 05:35:46.659230] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:13.358 [2024-12-07 05:35:46.661318] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:13.358 [2024-12-07 05:35:46.661352] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:13.358 BaseBdev1 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.358 BaseBdev2_malloc 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.358 true 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.358 [2024-12-07 05:35:46.699583] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:13.358 [2024-12-07 05:35:46.699646] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:13.358 [2024-12-07 05:35:46.699665] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:13.358 [2024-12-07 05:35:46.699681] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:13.358 [2024-12-07 05:35:46.701793] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:13.358 [2024-12-07 05:35:46.701827] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:13.358 BaseBdev2 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.358 BaseBdev3_malloc 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.358 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.618 true 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.618 [2024-12-07 05:35:46.740085] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:13.618 [2024-12-07 05:35:46.740130] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:13.618 [2024-12-07 05:35:46.740164] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:13.618 [2024-12-07 05:35:46.740172] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:13.618 [2024-12-07 05:35:46.742283] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:13.618 [2024-12-07 05:35:46.742366] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:13.618 BaseBdev3 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.618 [2024-12-07 05:35:46.752118] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:13.618 [2024-12-07 05:35:46.753933] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:13.618 [2024-12-07 05:35:46.754003] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:13.618 [2024-12-07 05:35:46.754172] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:13.618 [2024-12-07 05:35:46.754186] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:13.618 [2024-12-07 05:35:46.754428] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:08:13.618 [2024-12-07 05:35:46.754587] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:13.618 [2024-12-07 05:35:46.754603] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:13.618 [2024-12-07 05:35:46.754735] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:13.618 "name": "raid_bdev1", 00:08:13.618 "uuid": "45efb24c-5a09-4a22-b8d9-10eff76ffc23", 00:08:13.618 "strip_size_kb": 64, 00:08:13.618 "state": "online", 00:08:13.618 "raid_level": "raid0", 00:08:13.618 "superblock": true, 00:08:13.618 "num_base_bdevs": 3, 00:08:13.618 "num_base_bdevs_discovered": 3, 00:08:13.618 "num_base_bdevs_operational": 3, 00:08:13.618 "base_bdevs_list": [ 00:08:13.618 { 00:08:13.618 "name": "BaseBdev1", 00:08:13.618 "uuid": "29f3a95d-e3f9-56af-9b56-3600e980e3c4", 00:08:13.618 "is_configured": true, 00:08:13.618 "data_offset": 2048, 00:08:13.618 "data_size": 63488 00:08:13.618 }, 00:08:13.618 { 00:08:13.618 "name": "BaseBdev2", 00:08:13.618 "uuid": "d431290d-f988-55d6-bfe7-b31dc04fe4de", 00:08:13.618 "is_configured": true, 00:08:13.618 "data_offset": 2048, 00:08:13.618 "data_size": 63488 00:08:13.618 }, 00:08:13.618 { 00:08:13.618 "name": "BaseBdev3", 00:08:13.618 "uuid": "cf4e9816-2c16-59ba-8eaa-b0b054372f23", 00:08:13.618 "is_configured": true, 00:08:13.618 "data_offset": 2048, 00:08:13.618 "data_size": 63488 00:08:13.618 } 00:08:13.618 ] 00:08:13.618 }' 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:13.618 05:35:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.877 05:35:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:13.877 05:35:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:14.136 [2024-12-07 05:35:47.311522] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:08:15.073 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:15.073 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.074 "name": "raid_bdev1", 00:08:15.074 "uuid": "45efb24c-5a09-4a22-b8d9-10eff76ffc23", 00:08:15.074 "strip_size_kb": 64, 00:08:15.074 "state": "online", 00:08:15.074 "raid_level": "raid0", 00:08:15.074 "superblock": true, 00:08:15.074 "num_base_bdevs": 3, 00:08:15.074 "num_base_bdevs_discovered": 3, 00:08:15.074 "num_base_bdevs_operational": 3, 00:08:15.074 "base_bdevs_list": [ 00:08:15.074 { 00:08:15.074 "name": "BaseBdev1", 00:08:15.074 "uuid": "29f3a95d-e3f9-56af-9b56-3600e980e3c4", 00:08:15.074 "is_configured": true, 00:08:15.074 "data_offset": 2048, 00:08:15.074 "data_size": 63488 00:08:15.074 }, 00:08:15.074 { 00:08:15.074 "name": "BaseBdev2", 00:08:15.074 "uuid": "d431290d-f988-55d6-bfe7-b31dc04fe4de", 00:08:15.074 "is_configured": true, 00:08:15.074 "data_offset": 2048, 00:08:15.074 "data_size": 63488 00:08:15.074 }, 00:08:15.074 { 00:08:15.074 "name": "BaseBdev3", 00:08:15.074 "uuid": "cf4e9816-2c16-59ba-8eaa-b0b054372f23", 00:08:15.074 "is_configured": true, 00:08:15.074 "data_offset": 2048, 00:08:15.074 "data_size": 63488 00:08:15.074 } 00:08:15.074 ] 00:08:15.074 }' 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.074 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.333 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:15.333 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.333 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.333 [2024-12-07 05:35:48.643295] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:15.333 [2024-12-07 05:35:48.643326] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:15.333 [2024-12-07 05:35:48.645946] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:15.333 [2024-12-07 05:35:48.646002] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:15.333 [2024-12-07 05:35:48.646038] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:15.333 [2024-12-07 05:35:48.646050] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:15.333 { 00:08:15.333 "results": [ 00:08:15.333 { 00:08:15.333 "job": "raid_bdev1", 00:08:15.333 "core_mask": "0x1", 00:08:15.333 "workload": "randrw", 00:08:15.333 "percentage": 50, 00:08:15.333 "status": "finished", 00:08:15.333 "queue_depth": 1, 00:08:15.333 "io_size": 131072, 00:08:15.333 "runtime": 1.332526, 00:08:15.333 "iops": 16305.122751826231, 00:08:15.333 "mibps": 2038.140343978279, 00:08:15.333 "io_failed": 1, 00:08:15.333 "io_timeout": 0, 00:08:15.333 "avg_latency_us": 84.7955753066094, 00:08:15.333 "min_latency_us": 25.3764192139738, 00:08:15.333 "max_latency_us": 1402.2986899563318 00:08:15.333 } 00:08:15.333 ], 00:08:15.333 "core_count": 1 00:08:15.334 } 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76292 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 76292 ']' 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 76292 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76292 00:08:15.334 killing process with pid 76292 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76292' 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 76292 00:08:15.334 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 76292 00:08:15.334 [2024-12-07 05:35:48.692000] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:15.593 [2024-12-07 05:35:48.717966] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:15.593 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.JqNlcoDMXK 00:08:15.593 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:15.593 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:15.593 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.75 00:08:15.593 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:15.593 ************************************ 00:08:15.593 END TEST raid_write_error_test 00:08:15.593 ************************************ 00:08:15.593 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:15.593 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:15.593 05:35:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.75 != \0\.\0\0 ]] 00:08:15.593 00:08:15.593 real 0m3.225s 00:08:15.593 user 0m4.114s 00:08:15.593 sys 0m0.499s 00:08:15.593 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:15.593 05:35:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.853 05:35:48 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:15.853 05:35:48 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:08:15.853 05:35:48 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:15.853 05:35:48 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:15.853 05:35:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:15.853 ************************************ 00:08:15.853 START TEST raid_state_function_test 00:08:15.853 ************************************ 00:08:15.853 05:35:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 false 00:08:15.853 05:35:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:15.853 05:35:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:15.853 05:35:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:15.853 05:35:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:15.853 Process raid pid: 76419 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=76419 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 76419' 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 76419 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 76419 ']' 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:15.853 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:15.853 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.853 [2024-12-07 05:35:49.090620] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:08:15.853 [2024-12-07 05:35:49.090857] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:16.138 [2024-12-07 05:35:49.226701] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:16.138 [2024-12-07 05:35:49.251468] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:16.138 [2024-12-07 05:35:49.292972] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:16.138 [2024-12-07 05:35:49.293087] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.717 [2024-12-07 05:35:49.915130] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:16.717 [2024-12-07 05:35:49.915243] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:16.717 [2024-12-07 05:35:49.915288] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:16.717 [2024-12-07 05:35:49.915314] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:16.717 [2024-12-07 05:35:49.915358] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:16.717 [2024-12-07 05:35:49.915386] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:16.717 "name": "Existed_Raid", 00:08:16.717 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.717 "strip_size_kb": 64, 00:08:16.717 "state": "configuring", 00:08:16.717 "raid_level": "concat", 00:08:16.717 "superblock": false, 00:08:16.717 "num_base_bdevs": 3, 00:08:16.717 "num_base_bdevs_discovered": 0, 00:08:16.717 "num_base_bdevs_operational": 3, 00:08:16.717 "base_bdevs_list": [ 00:08:16.717 { 00:08:16.717 "name": "BaseBdev1", 00:08:16.717 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.717 "is_configured": false, 00:08:16.717 "data_offset": 0, 00:08:16.717 "data_size": 0 00:08:16.717 }, 00:08:16.717 { 00:08:16.717 "name": "BaseBdev2", 00:08:16.717 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.717 "is_configured": false, 00:08:16.717 "data_offset": 0, 00:08:16.717 "data_size": 0 00:08:16.717 }, 00:08:16.717 { 00:08:16.717 "name": "BaseBdev3", 00:08:16.717 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.717 "is_configured": false, 00:08:16.717 "data_offset": 0, 00:08:16.717 "data_size": 0 00:08:16.717 } 00:08:16.717 ] 00:08:16.717 }' 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:16.717 05:35:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.976 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:16.976 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.976 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.976 [2024-12-07 05:35:50.330349] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:16.976 [2024-12-07 05:35:50.330386] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:16.976 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.976 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:16.976 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.976 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.976 [2024-12-07 05:35:50.338359] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:16.976 [2024-12-07 05:35:50.338410] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:16.976 [2024-12-07 05:35:50.338419] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:16.976 [2024-12-07 05:35:50.338428] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:16.976 [2024-12-07 05:35:50.338434] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:16.976 [2024-12-07 05:35:50.338442] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.237 [2024-12-07 05:35:50.355070] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:17.237 BaseBdev1 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.237 [ 00:08:17.237 { 00:08:17.237 "name": "BaseBdev1", 00:08:17.237 "aliases": [ 00:08:17.237 "7c7d656b-71fa-47df-aa51-6401189e4aeb" 00:08:17.237 ], 00:08:17.237 "product_name": "Malloc disk", 00:08:17.237 "block_size": 512, 00:08:17.237 "num_blocks": 65536, 00:08:17.237 "uuid": "7c7d656b-71fa-47df-aa51-6401189e4aeb", 00:08:17.237 "assigned_rate_limits": { 00:08:17.237 "rw_ios_per_sec": 0, 00:08:17.237 "rw_mbytes_per_sec": 0, 00:08:17.237 "r_mbytes_per_sec": 0, 00:08:17.237 "w_mbytes_per_sec": 0 00:08:17.237 }, 00:08:17.237 "claimed": true, 00:08:17.237 "claim_type": "exclusive_write", 00:08:17.237 "zoned": false, 00:08:17.237 "supported_io_types": { 00:08:17.237 "read": true, 00:08:17.237 "write": true, 00:08:17.237 "unmap": true, 00:08:17.237 "flush": true, 00:08:17.237 "reset": true, 00:08:17.237 "nvme_admin": false, 00:08:17.237 "nvme_io": false, 00:08:17.237 "nvme_io_md": false, 00:08:17.237 "write_zeroes": true, 00:08:17.237 "zcopy": true, 00:08:17.237 "get_zone_info": false, 00:08:17.237 "zone_management": false, 00:08:17.237 "zone_append": false, 00:08:17.237 "compare": false, 00:08:17.237 "compare_and_write": false, 00:08:17.237 "abort": true, 00:08:17.237 "seek_hole": false, 00:08:17.237 "seek_data": false, 00:08:17.237 "copy": true, 00:08:17.237 "nvme_iov_md": false 00:08:17.237 }, 00:08:17.237 "memory_domains": [ 00:08:17.237 { 00:08:17.237 "dma_device_id": "system", 00:08:17.237 "dma_device_type": 1 00:08:17.237 }, 00:08:17.237 { 00:08:17.237 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:17.237 "dma_device_type": 2 00:08:17.237 } 00:08:17.237 ], 00:08:17.237 "driver_specific": {} 00:08:17.237 } 00:08:17.237 ] 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.237 "name": "Existed_Raid", 00:08:17.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.237 "strip_size_kb": 64, 00:08:17.237 "state": "configuring", 00:08:17.237 "raid_level": "concat", 00:08:17.237 "superblock": false, 00:08:17.237 "num_base_bdevs": 3, 00:08:17.237 "num_base_bdevs_discovered": 1, 00:08:17.237 "num_base_bdevs_operational": 3, 00:08:17.237 "base_bdevs_list": [ 00:08:17.237 { 00:08:17.237 "name": "BaseBdev1", 00:08:17.237 "uuid": "7c7d656b-71fa-47df-aa51-6401189e4aeb", 00:08:17.237 "is_configured": true, 00:08:17.237 "data_offset": 0, 00:08:17.237 "data_size": 65536 00:08:17.237 }, 00:08:17.237 { 00:08:17.237 "name": "BaseBdev2", 00:08:17.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.237 "is_configured": false, 00:08:17.237 "data_offset": 0, 00:08:17.237 "data_size": 0 00:08:17.237 }, 00:08:17.237 { 00:08:17.237 "name": "BaseBdev3", 00:08:17.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.237 "is_configured": false, 00:08:17.237 "data_offset": 0, 00:08:17.237 "data_size": 0 00:08:17.237 } 00:08:17.237 ] 00:08:17.237 }' 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.237 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.497 [2024-12-07 05:35:50.766426] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:17.497 [2024-12-07 05:35:50.766523] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.497 [2024-12-07 05:35:50.774463] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:17.497 [2024-12-07 05:35:50.776360] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:17.497 [2024-12-07 05:35:50.776437] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:17.497 [2024-12-07 05:35:50.776471] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:17.497 [2024-12-07 05:35:50.776512] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.497 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.497 "name": "Existed_Raid", 00:08:17.497 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.497 "strip_size_kb": 64, 00:08:17.497 "state": "configuring", 00:08:17.497 "raid_level": "concat", 00:08:17.497 "superblock": false, 00:08:17.497 "num_base_bdevs": 3, 00:08:17.498 "num_base_bdevs_discovered": 1, 00:08:17.498 "num_base_bdevs_operational": 3, 00:08:17.498 "base_bdevs_list": [ 00:08:17.498 { 00:08:17.498 "name": "BaseBdev1", 00:08:17.498 "uuid": "7c7d656b-71fa-47df-aa51-6401189e4aeb", 00:08:17.498 "is_configured": true, 00:08:17.498 "data_offset": 0, 00:08:17.498 "data_size": 65536 00:08:17.498 }, 00:08:17.498 { 00:08:17.498 "name": "BaseBdev2", 00:08:17.498 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.498 "is_configured": false, 00:08:17.498 "data_offset": 0, 00:08:17.498 "data_size": 0 00:08:17.498 }, 00:08:17.498 { 00:08:17.498 "name": "BaseBdev3", 00:08:17.498 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.498 "is_configured": false, 00:08:17.498 "data_offset": 0, 00:08:17.498 "data_size": 0 00:08:17.498 } 00:08:17.498 ] 00:08:17.498 }' 00:08:17.498 05:35:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.498 05:35:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.088 [2024-12-07 05:35:51.196478] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:18.088 BaseBdev2 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.088 [ 00:08:18.088 { 00:08:18.088 "name": "BaseBdev2", 00:08:18.088 "aliases": [ 00:08:18.088 "6b9abb78-2994-456a-ab90-b2c3e6f8be9a" 00:08:18.088 ], 00:08:18.088 "product_name": "Malloc disk", 00:08:18.088 "block_size": 512, 00:08:18.088 "num_blocks": 65536, 00:08:18.088 "uuid": "6b9abb78-2994-456a-ab90-b2c3e6f8be9a", 00:08:18.088 "assigned_rate_limits": { 00:08:18.088 "rw_ios_per_sec": 0, 00:08:18.088 "rw_mbytes_per_sec": 0, 00:08:18.088 "r_mbytes_per_sec": 0, 00:08:18.088 "w_mbytes_per_sec": 0 00:08:18.088 }, 00:08:18.088 "claimed": true, 00:08:18.088 "claim_type": "exclusive_write", 00:08:18.088 "zoned": false, 00:08:18.088 "supported_io_types": { 00:08:18.088 "read": true, 00:08:18.088 "write": true, 00:08:18.088 "unmap": true, 00:08:18.088 "flush": true, 00:08:18.088 "reset": true, 00:08:18.088 "nvme_admin": false, 00:08:18.088 "nvme_io": false, 00:08:18.088 "nvme_io_md": false, 00:08:18.088 "write_zeroes": true, 00:08:18.088 "zcopy": true, 00:08:18.088 "get_zone_info": false, 00:08:18.088 "zone_management": false, 00:08:18.088 "zone_append": false, 00:08:18.088 "compare": false, 00:08:18.088 "compare_and_write": false, 00:08:18.088 "abort": true, 00:08:18.088 "seek_hole": false, 00:08:18.088 "seek_data": false, 00:08:18.088 "copy": true, 00:08:18.088 "nvme_iov_md": false 00:08:18.088 }, 00:08:18.088 "memory_domains": [ 00:08:18.088 { 00:08:18.088 "dma_device_id": "system", 00:08:18.088 "dma_device_type": 1 00:08:18.088 }, 00:08:18.088 { 00:08:18.088 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.088 "dma_device_type": 2 00:08:18.088 } 00:08:18.088 ], 00:08:18.088 "driver_specific": {} 00:08:18.088 } 00:08:18.088 ] 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.088 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.089 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:18.089 "name": "Existed_Raid", 00:08:18.089 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:18.089 "strip_size_kb": 64, 00:08:18.089 "state": "configuring", 00:08:18.089 "raid_level": "concat", 00:08:18.089 "superblock": false, 00:08:18.089 "num_base_bdevs": 3, 00:08:18.089 "num_base_bdevs_discovered": 2, 00:08:18.089 "num_base_bdevs_operational": 3, 00:08:18.089 "base_bdevs_list": [ 00:08:18.089 { 00:08:18.089 "name": "BaseBdev1", 00:08:18.089 "uuid": "7c7d656b-71fa-47df-aa51-6401189e4aeb", 00:08:18.089 "is_configured": true, 00:08:18.089 "data_offset": 0, 00:08:18.089 "data_size": 65536 00:08:18.089 }, 00:08:18.089 { 00:08:18.089 "name": "BaseBdev2", 00:08:18.089 "uuid": "6b9abb78-2994-456a-ab90-b2c3e6f8be9a", 00:08:18.089 "is_configured": true, 00:08:18.089 "data_offset": 0, 00:08:18.089 "data_size": 65536 00:08:18.089 }, 00:08:18.089 { 00:08:18.089 "name": "BaseBdev3", 00:08:18.089 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:18.089 "is_configured": false, 00:08:18.089 "data_offset": 0, 00:08:18.089 "data_size": 0 00:08:18.089 } 00:08:18.089 ] 00:08:18.089 }' 00:08:18.089 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:18.089 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.349 [2024-12-07 05:35:51.679157] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:18.349 [2024-12-07 05:35:51.679437] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:18.349 [2024-12-07 05:35:51.679548] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:18.349 [2024-12-07 05:35:51.680712] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:18.349 [2024-12-07 05:35:51.681364] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:18.349 [2024-12-07 05:35:51.681421] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:18.349 [2024-12-07 05:35:51.682094] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:18.349 BaseBdev3 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.349 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.349 [ 00:08:18.349 { 00:08:18.349 "name": "BaseBdev3", 00:08:18.349 "aliases": [ 00:08:18.349 "4c8dd6df-6e62-42c4-9bda-8f7204eaec10" 00:08:18.349 ], 00:08:18.349 "product_name": "Malloc disk", 00:08:18.349 "block_size": 512, 00:08:18.349 "num_blocks": 65536, 00:08:18.349 "uuid": "4c8dd6df-6e62-42c4-9bda-8f7204eaec10", 00:08:18.349 "assigned_rate_limits": { 00:08:18.349 "rw_ios_per_sec": 0, 00:08:18.349 "rw_mbytes_per_sec": 0, 00:08:18.349 "r_mbytes_per_sec": 0, 00:08:18.349 "w_mbytes_per_sec": 0 00:08:18.349 }, 00:08:18.349 "claimed": true, 00:08:18.349 "claim_type": "exclusive_write", 00:08:18.349 "zoned": false, 00:08:18.349 "supported_io_types": { 00:08:18.349 "read": true, 00:08:18.349 "write": true, 00:08:18.349 "unmap": true, 00:08:18.349 "flush": true, 00:08:18.609 "reset": true, 00:08:18.609 "nvme_admin": false, 00:08:18.609 "nvme_io": false, 00:08:18.609 "nvme_io_md": false, 00:08:18.609 "write_zeroes": true, 00:08:18.609 "zcopy": true, 00:08:18.609 "get_zone_info": false, 00:08:18.609 "zone_management": false, 00:08:18.609 "zone_append": false, 00:08:18.609 "compare": false, 00:08:18.609 "compare_and_write": false, 00:08:18.609 "abort": true, 00:08:18.609 "seek_hole": false, 00:08:18.609 "seek_data": false, 00:08:18.609 "copy": true, 00:08:18.609 "nvme_iov_md": false 00:08:18.609 }, 00:08:18.609 "memory_domains": [ 00:08:18.609 { 00:08:18.609 "dma_device_id": "system", 00:08:18.609 "dma_device_type": 1 00:08:18.609 }, 00:08:18.609 { 00:08:18.609 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.609 "dma_device_type": 2 00:08:18.609 } 00:08:18.609 ], 00:08:18.609 "driver_specific": {} 00:08:18.609 } 00:08:18.609 ] 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:18.609 "name": "Existed_Raid", 00:08:18.609 "uuid": "3183e0ec-f941-427b-94ed-8606773e2d2d", 00:08:18.609 "strip_size_kb": 64, 00:08:18.609 "state": "online", 00:08:18.609 "raid_level": "concat", 00:08:18.609 "superblock": false, 00:08:18.609 "num_base_bdevs": 3, 00:08:18.609 "num_base_bdevs_discovered": 3, 00:08:18.609 "num_base_bdevs_operational": 3, 00:08:18.609 "base_bdevs_list": [ 00:08:18.609 { 00:08:18.609 "name": "BaseBdev1", 00:08:18.609 "uuid": "7c7d656b-71fa-47df-aa51-6401189e4aeb", 00:08:18.609 "is_configured": true, 00:08:18.609 "data_offset": 0, 00:08:18.609 "data_size": 65536 00:08:18.609 }, 00:08:18.609 { 00:08:18.609 "name": "BaseBdev2", 00:08:18.609 "uuid": "6b9abb78-2994-456a-ab90-b2c3e6f8be9a", 00:08:18.609 "is_configured": true, 00:08:18.609 "data_offset": 0, 00:08:18.609 "data_size": 65536 00:08:18.609 }, 00:08:18.609 { 00:08:18.609 "name": "BaseBdev3", 00:08:18.609 "uuid": "4c8dd6df-6e62-42c4-9bda-8f7204eaec10", 00:08:18.609 "is_configured": true, 00:08:18.609 "data_offset": 0, 00:08:18.609 "data_size": 65536 00:08:18.609 } 00:08:18.609 ] 00:08:18.609 }' 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:18.609 05:35:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:18.869 [2024-12-07 05:35:52.174635] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:18.869 "name": "Existed_Raid", 00:08:18.869 "aliases": [ 00:08:18.869 "3183e0ec-f941-427b-94ed-8606773e2d2d" 00:08:18.869 ], 00:08:18.869 "product_name": "Raid Volume", 00:08:18.869 "block_size": 512, 00:08:18.869 "num_blocks": 196608, 00:08:18.869 "uuid": "3183e0ec-f941-427b-94ed-8606773e2d2d", 00:08:18.869 "assigned_rate_limits": { 00:08:18.869 "rw_ios_per_sec": 0, 00:08:18.869 "rw_mbytes_per_sec": 0, 00:08:18.869 "r_mbytes_per_sec": 0, 00:08:18.869 "w_mbytes_per_sec": 0 00:08:18.869 }, 00:08:18.869 "claimed": false, 00:08:18.869 "zoned": false, 00:08:18.869 "supported_io_types": { 00:08:18.869 "read": true, 00:08:18.869 "write": true, 00:08:18.869 "unmap": true, 00:08:18.869 "flush": true, 00:08:18.869 "reset": true, 00:08:18.869 "nvme_admin": false, 00:08:18.869 "nvme_io": false, 00:08:18.869 "nvme_io_md": false, 00:08:18.869 "write_zeroes": true, 00:08:18.869 "zcopy": false, 00:08:18.869 "get_zone_info": false, 00:08:18.869 "zone_management": false, 00:08:18.869 "zone_append": false, 00:08:18.869 "compare": false, 00:08:18.869 "compare_and_write": false, 00:08:18.869 "abort": false, 00:08:18.869 "seek_hole": false, 00:08:18.869 "seek_data": false, 00:08:18.869 "copy": false, 00:08:18.869 "nvme_iov_md": false 00:08:18.869 }, 00:08:18.869 "memory_domains": [ 00:08:18.869 { 00:08:18.869 "dma_device_id": "system", 00:08:18.869 "dma_device_type": 1 00:08:18.869 }, 00:08:18.869 { 00:08:18.869 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.869 "dma_device_type": 2 00:08:18.869 }, 00:08:18.869 { 00:08:18.869 "dma_device_id": "system", 00:08:18.869 "dma_device_type": 1 00:08:18.869 }, 00:08:18.869 { 00:08:18.869 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.869 "dma_device_type": 2 00:08:18.869 }, 00:08:18.869 { 00:08:18.869 "dma_device_id": "system", 00:08:18.869 "dma_device_type": 1 00:08:18.869 }, 00:08:18.869 { 00:08:18.869 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.869 "dma_device_type": 2 00:08:18.869 } 00:08:18.869 ], 00:08:18.869 "driver_specific": { 00:08:18.869 "raid": { 00:08:18.869 "uuid": "3183e0ec-f941-427b-94ed-8606773e2d2d", 00:08:18.869 "strip_size_kb": 64, 00:08:18.869 "state": "online", 00:08:18.869 "raid_level": "concat", 00:08:18.869 "superblock": false, 00:08:18.869 "num_base_bdevs": 3, 00:08:18.869 "num_base_bdevs_discovered": 3, 00:08:18.869 "num_base_bdevs_operational": 3, 00:08:18.869 "base_bdevs_list": [ 00:08:18.869 { 00:08:18.869 "name": "BaseBdev1", 00:08:18.869 "uuid": "7c7d656b-71fa-47df-aa51-6401189e4aeb", 00:08:18.869 "is_configured": true, 00:08:18.869 "data_offset": 0, 00:08:18.869 "data_size": 65536 00:08:18.869 }, 00:08:18.869 { 00:08:18.869 "name": "BaseBdev2", 00:08:18.869 "uuid": "6b9abb78-2994-456a-ab90-b2c3e6f8be9a", 00:08:18.869 "is_configured": true, 00:08:18.869 "data_offset": 0, 00:08:18.869 "data_size": 65536 00:08:18.869 }, 00:08:18.869 { 00:08:18.869 "name": "BaseBdev3", 00:08:18.869 "uuid": "4c8dd6df-6e62-42c4-9bda-8f7204eaec10", 00:08:18.869 "is_configured": true, 00:08:18.869 "data_offset": 0, 00:08:18.869 "data_size": 65536 00:08:18.869 } 00:08:18.869 ] 00:08:18.869 } 00:08:18.869 } 00:08:18.869 }' 00:08:18.869 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:19.128 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:19.128 BaseBdev2 00:08:19.128 BaseBdev3' 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.129 [2024-12-07 05:35:52.465831] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:19.129 [2024-12-07 05:35:52.465859] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:19.129 [2024-12-07 05:35:52.465926] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.129 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.387 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.387 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:19.387 "name": "Existed_Raid", 00:08:19.387 "uuid": "3183e0ec-f941-427b-94ed-8606773e2d2d", 00:08:19.387 "strip_size_kb": 64, 00:08:19.387 "state": "offline", 00:08:19.387 "raid_level": "concat", 00:08:19.387 "superblock": false, 00:08:19.387 "num_base_bdevs": 3, 00:08:19.387 "num_base_bdevs_discovered": 2, 00:08:19.387 "num_base_bdevs_operational": 2, 00:08:19.387 "base_bdevs_list": [ 00:08:19.387 { 00:08:19.387 "name": null, 00:08:19.387 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:19.387 "is_configured": false, 00:08:19.387 "data_offset": 0, 00:08:19.387 "data_size": 65536 00:08:19.387 }, 00:08:19.387 { 00:08:19.387 "name": "BaseBdev2", 00:08:19.387 "uuid": "6b9abb78-2994-456a-ab90-b2c3e6f8be9a", 00:08:19.387 "is_configured": true, 00:08:19.387 "data_offset": 0, 00:08:19.387 "data_size": 65536 00:08:19.387 }, 00:08:19.387 { 00:08:19.387 "name": "BaseBdev3", 00:08:19.387 "uuid": "4c8dd6df-6e62-42c4-9bda-8f7204eaec10", 00:08:19.387 "is_configured": true, 00:08:19.387 "data_offset": 0, 00:08:19.387 "data_size": 65536 00:08:19.387 } 00:08:19.387 ] 00:08:19.387 }' 00:08:19.387 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:19.387 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.643 [2024-12-07 05:35:52.908536] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.643 [2024-12-07 05:35:52.959773] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:19.643 [2024-12-07 05:35:52.959823] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.643 05:35:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.901 BaseBdev2 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.901 [ 00:08:19.901 { 00:08:19.901 "name": "BaseBdev2", 00:08:19.901 "aliases": [ 00:08:19.901 "ebec53cd-45f9-4df2-9fde-7d292a85a0de" 00:08:19.901 ], 00:08:19.901 "product_name": "Malloc disk", 00:08:19.901 "block_size": 512, 00:08:19.901 "num_blocks": 65536, 00:08:19.901 "uuid": "ebec53cd-45f9-4df2-9fde-7d292a85a0de", 00:08:19.901 "assigned_rate_limits": { 00:08:19.901 "rw_ios_per_sec": 0, 00:08:19.901 "rw_mbytes_per_sec": 0, 00:08:19.901 "r_mbytes_per_sec": 0, 00:08:19.901 "w_mbytes_per_sec": 0 00:08:19.901 }, 00:08:19.901 "claimed": false, 00:08:19.901 "zoned": false, 00:08:19.901 "supported_io_types": { 00:08:19.901 "read": true, 00:08:19.901 "write": true, 00:08:19.901 "unmap": true, 00:08:19.901 "flush": true, 00:08:19.901 "reset": true, 00:08:19.901 "nvme_admin": false, 00:08:19.901 "nvme_io": false, 00:08:19.901 "nvme_io_md": false, 00:08:19.901 "write_zeroes": true, 00:08:19.901 "zcopy": true, 00:08:19.901 "get_zone_info": false, 00:08:19.901 "zone_management": false, 00:08:19.901 "zone_append": false, 00:08:19.901 "compare": false, 00:08:19.901 "compare_and_write": false, 00:08:19.901 "abort": true, 00:08:19.901 "seek_hole": false, 00:08:19.901 "seek_data": false, 00:08:19.901 "copy": true, 00:08:19.901 "nvme_iov_md": false 00:08:19.901 }, 00:08:19.901 "memory_domains": [ 00:08:19.901 { 00:08:19.901 "dma_device_id": "system", 00:08:19.901 "dma_device_type": 1 00:08:19.901 }, 00:08:19.901 { 00:08:19.901 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:19.901 "dma_device_type": 2 00:08:19.901 } 00:08:19.901 ], 00:08:19.901 "driver_specific": {} 00:08:19.901 } 00:08:19.901 ] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.901 BaseBdev3 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.901 [ 00:08:19.901 { 00:08:19.901 "name": "BaseBdev3", 00:08:19.901 "aliases": [ 00:08:19.901 "5c2cd370-3a48-4ab2-94f3-cf002f9c9832" 00:08:19.901 ], 00:08:19.901 "product_name": "Malloc disk", 00:08:19.901 "block_size": 512, 00:08:19.901 "num_blocks": 65536, 00:08:19.901 "uuid": "5c2cd370-3a48-4ab2-94f3-cf002f9c9832", 00:08:19.901 "assigned_rate_limits": { 00:08:19.901 "rw_ios_per_sec": 0, 00:08:19.901 "rw_mbytes_per_sec": 0, 00:08:19.901 "r_mbytes_per_sec": 0, 00:08:19.901 "w_mbytes_per_sec": 0 00:08:19.901 }, 00:08:19.901 "claimed": false, 00:08:19.901 "zoned": false, 00:08:19.901 "supported_io_types": { 00:08:19.901 "read": true, 00:08:19.901 "write": true, 00:08:19.901 "unmap": true, 00:08:19.901 "flush": true, 00:08:19.901 "reset": true, 00:08:19.901 "nvme_admin": false, 00:08:19.901 "nvme_io": false, 00:08:19.901 "nvme_io_md": false, 00:08:19.901 "write_zeroes": true, 00:08:19.901 "zcopy": true, 00:08:19.901 "get_zone_info": false, 00:08:19.901 "zone_management": false, 00:08:19.901 "zone_append": false, 00:08:19.901 "compare": false, 00:08:19.901 "compare_and_write": false, 00:08:19.901 "abort": true, 00:08:19.901 "seek_hole": false, 00:08:19.901 "seek_data": false, 00:08:19.901 "copy": true, 00:08:19.901 "nvme_iov_md": false 00:08:19.901 }, 00:08:19.901 "memory_domains": [ 00:08:19.901 { 00:08:19.901 "dma_device_id": "system", 00:08:19.901 "dma_device_type": 1 00:08:19.901 }, 00:08:19.901 { 00:08:19.901 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:19.901 "dma_device_type": 2 00:08:19.901 } 00:08:19.901 ], 00:08:19.901 "driver_specific": {} 00:08:19.901 } 00:08:19.901 ] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.901 [2024-12-07 05:35:53.115067] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:19.901 [2024-12-07 05:35:53.115167] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:19.901 [2024-12-07 05:35:53.115211] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:19.901 [2024-12-07 05:35:53.116982] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:19.901 "name": "Existed_Raid", 00:08:19.901 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:19.901 "strip_size_kb": 64, 00:08:19.901 "state": "configuring", 00:08:19.901 "raid_level": "concat", 00:08:19.901 "superblock": false, 00:08:19.901 "num_base_bdevs": 3, 00:08:19.901 "num_base_bdevs_discovered": 2, 00:08:19.901 "num_base_bdevs_operational": 3, 00:08:19.901 "base_bdevs_list": [ 00:08:19.901 { 00:08:19.901 "name": "BaseBdev1", 00:08:19.901 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:19.901 "is_configured": false, 00:08:19.901 "data_offset": 0, 00:08:19.901 "data_size": 0 00:08:19.901 }, 00:08:19.901 { 00:08:19.901 "name": "BaseBdev2", 00:08:19.901 "uuid": "ebec53cd-45f9-4df2-9fde-7d292a85a0de", 00:08:19.901 "is_configured": true, 00:08:19.901 "data_offset": 0, 00:08:19.901 "data_size": 65536 00:08:19.901 }, 00:08:19.901 { 00:08:19.901 "name": "BaseBdev3", 00:08:19.901 "uuid": "5c2cd370-3a48-4ab2-94f3-cf002f9c9832", 00:08:19.901 "is_configured": true, 00:08:19.901 "data_offset": 0, 00:08:19.901 "data_size": 65536 00:08:19.901 } 00:08:19.901 ] 00:08:19.901 }' 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:19.901 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.468 [2024-12-07 05:35:53.542527] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.468 "name": "Existed_Raid", 00:08:20.468 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.468 "strip_size_kb": 64, 00:08:20.468 "state": "configuring", 00:08:20.468 "raid_level": "concat", 00:08:20.468 "superblock": false, 00:08:20.468 "num_base_bdevs": 3, 00:08:20.468 "num_base_bdevs_discovered": 1, 00:08:20.468 "num_base_bdevs_operational": 3, 00:08:20.468 "base_bdevs_list": [ 00:08:20.468 { 00:08:20.468 "name": "BaseBdev1", 00:08:20.468 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.468 "is_configured": false, 00:08:20.468 "data_offset": 0, 00:08:20.468 "data_size": 0 00:08:20.468 }, 00:08:20.468 { 00:08:20.468 "name": null, 00:08:20.468 "uuid": "ebec53cd-45f9-4df2-9fde-7d292a85a0de", 00:08:20.468 "is_configured": false, 00:08:20.468 "data_offset": 0, 00:08:20.468 "data_size": 65536 00:08:20.468 }, 00:08:20.468 { 00:08:20.468 "name": "BaseBdev3", 00:08:20.468 "uuid": "5c2cd370-3a48-4ab2-94f3-cf002f9c9832", 00:08:20.468 "is_configured": true, 00:08:20.468 "data_offset": 0, 00:08:20.468 "data_size": 65536 00:08:20.468 } 00:08:20.468 ] 00:08:20.468 }' 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.468 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.727 [2024-12-07 05:35:53.976641] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:20.727 BaseBdev1 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.727 05:35:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.727 [ 00:08:20.727 { 00:08:20.727 "name": "BaseBdev1", 00:08:20.727 "aliases": [ 00:08:20.727 "500dce35-dd74-414f-a9c7-73a73b654232" 00:08:20.727 ], 00:08:20.727 "product_name": "Malloc disk", 00:08:20.727 "block_size": 512, 00:08:20.727 "num_blocks": 65536, 00:08:20.727 "uuid": "500dce35-dd74-414f-a9c7-73a73b654232", 00:08:20.727 "assigned_rate_limits": { 00:08:20.727 "rw_ios_per_sec": 0, 00:08:20.727 "rw_mbytes_per_sec": 0, 00:08:20.727 "r_mbytes_per_sec": 0, 00:08:20.727 "w_mbytes_per_sec": 0 00:08:20.727 }, 00:08:20.727 "claimed": true, 00:08:20.727 "claim_type": "exclusive_write", 00:08:20.727 "zoned": false, 00:08:20.727 "supported_io_types": { 00:08:20.727 "read": true, 00:08:20.727 "write": true, 00:08:20.727 "unmap": true, 00:08:20.727 "flush": true, 00:08:20.727 "reset": true, 00:08:20.727 "nvme_admin": false, 00:08:20.727 "nvme_io": false, 00:08:20.727 "nvme_io_md": false, 00:08:20.727 "write_zeroes": true, 00:08:20.727 "zcopy": true, 00:08:20.727 "get_zone_info": false, 00:08:20.727 "zone_management": false, 00:08:20.727 "zone_append": false, 00:08:20.727 "compare": false, 00:08:20.727 "compare_and_write": false, 00:08:20.727 "abort": true, 00:08:20.727 "seek_hole": false, 00:08:20.727 "seek_data": false, 00:08:20.727 "copy": true, 00:08:20.727 "nvme_iov_md": false 00:08:20.727 }, 00:08:20.727 "memory_domains": [ 00:08:20.727 { 00:08:20.727 "dma_device_id": "system", 00:08:20.727 "dma_device_type": 1 00:08:20.727 }, 00:08:20.727 { 00:08:20.727 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:20.727 "dma_device_type": 2 00:08:20.727 } 00:08:20.727 ], 00:08:20.727 "driver_specific": {} 00:08:20.727 } 00:08:20.727 ] 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.727 "name": "Existed_Raid", 00:08:20.727 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.727 "strip_size_kb": 64, 00:08:20.727 "state": "configuring", 00:08:20.727 "raid_level": "concat", 00:08:20.727 "superblock": false, 00:08:20.727 "num_base_bdevs": 3, 00:08:20.727 "num_base_bdevs_discovered": 2, 00:08:20.727 "num_base_bdevs_operational": 3, 00:08:20.727 "base_bdevs_list": [ 00:08:20.727 { 00:08:20.727 "name": "BaseBdev1", 00:08:20.727 "uuid": "500dce35-dd74-414f-a9c7-73a73b654232", 00:08:20.727 "is_configured": true, 00:08:20.727 "data_offset": 0, 00:08:20.727 "data_size": 65536 00:08:20.727 }, 00:08:20.727 { 00:08:20.727 "name": null, 00:08:20.727 "uuid": "ebec53cd-45f9-4df2-9fde-7d292a85a0de", 00:08:20.727 "is_configured": false, 00:08:20.727 "data_offset": 0, 00:08:20.727 "data_size": 65536 00:08:20.727 }, 00:08:20.727 { 00:08:20.727 "name": "BaseBdev3", 00:08:20.727 "uuid": "5c2cd370-3a48-4ab2-94f3-cf002f9c9832", 00:08:20.727 "is_configured": true, 00:08:20.727 "data_offset": 0, 00:08:20.727 "data_size": 65536 00:08:20.727 } 00:08:20.727 ] 00:08:20.727 }' 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.727 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.986 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.986 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.986 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.986 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.244 [2024-12-07 05:35:54.395957] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:21.244 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.245 "name": "Existed_Raid", 00:08:21.245 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.245 "strip_size_kb": 64, 00:08:21.245 "state": "configuring", 00:08:21.245 "raid_level": "concat", 00:08:21.245 "superblock": false, 00:08:21.245 "num_base_bdevs": 3, 00:08:21.245 "num_base_bdevs_discovered": 1, 00:08:21.245 "num_base_bdevs_operational": 3, 00:08:21.245 "base_bdevs_list": [ 00:08:21.245 { 00:08:21.245 "name": "BaseBdev1", 00:08:21.245 "uuid": "500dce35-dd74-414f-a9c7-73a73b654232", 00:08:21.245 "is_configured": true, 00:08:21.245 "data_offset": 0, 00:08:21.245 "data_size": 65536 00:08:21.245 }, 00:08:21.245 { 00:08:21.245 "name": null, 00:08:21.245 "uuid": "ebec53cd-45f9-4df2-9fde-7d292a85a0de", 00:08:21.245 "is_configured": false, 00:08:21.245 "data_offset": 0, 00:08:21.245 "data_size": 65536 00:08:21.245 }, 00:08:21.245 { 00:08:21.245 "name": null, 00:08:21.245 "uuid": "5c2cd370-3a48-4ab2-94f3-cf002f9c9832", 00:08:21.245 "is_configured": false, 00:08:21.245 "data_offset": 0, 00:08:21.245 "data_size": 65536 00:08:21.245 } 00:08:21.245 ] 00:08:21.245 }' 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.245 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.503 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:21.503 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.503 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.503 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.762 [2024-12-07 05:35:54.895164] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.762 "name": "Existed_Raid", 00:08:21.762 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.762 "strip_size_kb": 64, 00:08:21.762 "state": "configuring", 00:08:21.762 "raid_level": "concat", 00:08:21.762 "superblock": false, 00:08:21.762 "num_base_bdevs": 3, 00:08:21.762 "num_base_bdevs_discovered": 2, 00:08:21.762 "num_base_bdevs_operational": 3, 00:08:21.762 "base_bdevs_list": [ 00:08:21.762 { 00:08:21.762 "name": "BaseBdev1", 00:08:21.762 "uuid": "500dce35-dd74-414f-a9c7-73a73b654232", 00:08:21.762 "is_configured": true, 00:08:21.762 "data_offset": 0, 00:08:21.762 "data_size": 65536 00:08:21.762 }, 00:08:21.762 { 00:08:21.762 "name": null, 00:08:21.762 "uuid": "ebec53cd-45f9-4df2-9fde-7d292a85a0de", 00:08:21.762 "is_configured": false, 00:08:21.762 "data_offset": 0, 00:08:21.762 "data_size": 65536 00:08:21.762 }, 00:08:21.762 { 00:08:21.762 "name": "BaseBdev3", 00:08:21.762 "uuid": "5c2cd370-3a48-4ab2-94f3-cf002f9c9832", 00:08:21.762 "is_configured": true, 00:08:21.762 "data_offset": 0, 00:08:21.762 "data_size": 65536 00:08:21.762 } 00:08:21.762 ] 00:08:21.762 }' 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.762 05:35:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.019 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.019 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:22.019 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.019 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.019 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.019 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:22.019 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:22.019 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.019 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.019 [2024-12-07 05:35:55.366569] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:22.019 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.020 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:22.276 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.277 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:22.277 "name": "Existed_Raid", 00:08:22.277 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.277 "strip_size_kb": 64, 00:08:22.277 "state": "configuring", 00:08:22.277 "raid_level": "concat", 00:08:22.277 "superblock": false, 00:08:22.277 "num_base_bdevs": 3, 00:08:22.277 "num_base_bdevs_discovered": 1, 00:08:22.277 "num_base_bdevs_operational": 3, 00:08:22.277 "base_bdevs_list": [ 00:08:22.277 { 00:08:22.277 "name": null, 00:08:22.277 "uuid": "500dce35-dd74-414f-a9c7-73a73b654232", 00:08:22.277 "is_configured": false, 00:08:22.277 "data_offset": 0, 00:08:22.277 "data_size": 65536 00:08:22.277 }, 00:08:22.277 { 00:08:22.277 "name": null, 00:08:22.277 "uuid": "ebec53cd-45f9-4df2-9fde-7d292a85a0de", 00:08:22.277 "is_configured": false, 00:08:22.277 "data_offset": 0, 00:08:22.277 "data_size": 65536 00:08:22.277 }, 00:08:22.277 { 00:08:22.277 "name": "BaseBdev3", 00:08:22.277 "uuid": "5c2cd370-3a48-4ab2-94f3-cf002f9c9832", 00:08:22.277 "is_configured": true, 00:08:22.277 "data_offset": 0, 00:08:22.277 "data_size": 65536 00:08:22.277 } 00:08:22.277 ] 00:08:22.277 }' 00:08:22.277 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:22.277 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.534 [2024-12-07 05:35:55.840350] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:22.534 "name": "Existed_Raid", 00:08:22.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.534 "strip_size_kb": 64, 00:08:22.534 "state": "configuring", 00:08:22.534 "raid_level": "concat", 00:08:22.534 "superblock": false, 00:08:22.534 "num_base_bdevs": 3, 00:08:22.534 "num_base_bdevs_discovered": 2, 00:08:22.534 "num_base_bdevs_operational": 3, 00:08:22.534 "base_bdevs_list": [ 00:08:22.534 { 00:08:22.534 "name": null, 00:08:22.534 "uuid": "500dce35-dd74-414f-a9c7-73a73b654232", 00:08:22.534 "is_configured": false, 00:08:22.534 "data_offset": 0, 00:08:22.534 "data_size": 65536 00:08:22.534 }, 00:08:22.534 { 00:08:22.534 "name": "BaseBdev2", 00:08:22.534 "uuid": "ebec53cd-45f9-4df2-9fde-7d292a85a0de", 00:08:22.534 "is_configured": true, 00:08:22.534 "data_offset": 0, 00:08:22.534 "data_size": 65536 00:08:22.534 }, 00:08:22.534 { 00:08:22.534 "name": "BaseBdev3", 00:08:22.534 "uuid": "5c2cd370-3a48-4ab2-94f3-cf002f9c9832", 00:08:22.534 "is_configured": true, 00:08:22.534 "data_offset": 0, 00:08:22.534 "data_size": 65536 00:08:22.534 } 00:08:22.534 ] 00:08:22.534 }' 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:22.534 05:35:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.099 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 500dce35-dd74-414f-a9c7-73a73b654232 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.100 [2024-12-07 05:35:56.366307] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:23.100 [2024-12-07 05:35:56.366418] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:23.100 [2024-12-07 05:35:56.366435] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:23.100 [2024-12-07 05:35:56.366689] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:23.100 [2024-12-07 05:35:56.366817] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:23.100 [2024-12-07 05:35:56.366827] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:23.100 [2024-12-07 05:35:56.367010] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:23.100 NewBaseBdev 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.100 [ 00:08:23.100 { 00:08:23.100 "name": "NewBaseBdev", 00:08:23.100 "aliases": [ 00:08:23.100 "500dce35-dd74-414f-a9c7-73a73b654232" 00:08:23.100 ], 00:08:23.100 "product_name": "Malloc disk", 00:08:23.100 "block_size": 512, 00:08:23.100 "num_blocks": 65536, 00:08:23.100 "uuid": "500dce35-dd74-414f-a9c7-73a73b654232", 00:08:23.100 "assigned_rate_limits": { 00:08:23.100 "rw_ios_per_sec": 0, 00:08:23.100 "rw_mbytes_per_sec": 0, 00:08:23.100 "r_mbytes_per_sec": 0, 00:08:23.100 "w_mbytes_per_sec": 0 00:08:23.100 }, 00:08:23.100 "claimed": true, 00:08:23.100 "claim_type": "exclusive_write", 00:08:23.100 "zoned": false, 00:08:23.100 "supported_io_types": { 00:08:23.100 "read": true, 00:08:23.100 "write": true, 00:08:23.100 "unmap": true, 00:08:23.100 "flush": true, 00:08:23.100 "reset": true, 00:08:23.100 "nvme_admin": false, 00:08:23.100 "nvme_io": false, 00:08:23.100 "nvme_io_md": false, 00:08:23.100 "write_zeroes": true, 00:08:23.100 "zcopy": true, 00:08:23.100 "get_zone_info": false, 00:08:23.100 "zone_management": false, 00:08:23.100 "zone_append": false, 00:08:23.100 "compare": false, 00:08:23.100 "compare_and_write": false, 00:08:23.100 "abort": true, 00:08:23.100 "seek_hole": false, 00:08:23.100 "seek_data": false, 00:08:23.100 "copy": true, 00:08:23.100 "nvme_iov_md": false 00:08:23.100 }, 00:08:23.100 "memory_domains": [ 00:08:23.100 { 00:08:23.100 "dma_device_id": "system", 00:08:23.100 "dma_device_type": 1 00:08:23.100 }, 00:08:23.100 { 00:08:23.100 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.100 "dma_device_type": 2 00:08:23.100 } 00:08:23.100 ], 00:08:23.100 "driver_specific": {} 00:08:23.100 } 00:08:23.100 ] 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:23.100 "name": "Existed_Raid", 00:08:23.100 "uuid": "df78b5a1-c9b2-4ca1-9de5-776dce56606f", 00:08:23.100 "strip_size_kb": 64, 00:08:23.100 "state": "online", 00:08:23.100 "raid_level": "concat", 00:08:23.100 "superblock": false, 00:08:23.100 "num_base_bdevs": 3, 00:08:23.100 "num_base_bdevs_discovered": 3, 00:08:23.100 "num_base_bdevs_operational": 3, 00:08:23.100 "base_bdevs_list": [ 00:08:23.100 { 00:08:23.100 "name": "NewBaseBdev", 00:08:23.100 "uuid": "500dce35-dd74-414f-a9c7-73a73b654232", 00:08:23.100 "is_configured": true, 00:08:23.100 "data_offset": 0, 00:08:23.100 "data_size": 65536 00:08:23.100 }, 00:08:23.100 { 00:08:23.100 "name": "BaseBdev2", 00:08:23.100 "uuid": "ebec53cd-45f9-4df2-9fde-7d292a85a0de", 00:08:23.100 "is_configured": true, 00:08:23.100 "data_offset": 0, 00:08:23.100 "data_size": 65536 00:08:23.100 }, 00:08:23.100 { 00:08:23.100 "name": "BaseBdev3", 00:08:23.100 "uuid": "5c2cd370-3a48-4ab2-94f3-cf002f9c9832", 00:08:23.100 "is_configured": true, 00:08:23.100 "data_offset": 0, 00:08:23.100 "data_size": 65536 00:08:23.100 } 00:08:23.100 ] 00:08:23.100 }' 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:23.100 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:23.668 [2024-12-07 05:35:56.805908] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:23.668 "name": "Existed_Raid", 00:08:23.668 "aliases": [ 00:08:23.668 "df78b5a1-c9b2-4ca1-9de5-776dce56606f" 00:08:23.668 ], 00:08:23.668 "product_name": "Raid Volume", 00:08:23.668 "block_size": 512, 00:08:23.668 "num_blocks": 196608, 00:08:23.668 "uuid": "df78b5a1-c9b2-4ca1-9de5-776dce56606f", 00:08:23.668 "assigned_rate_limits": { 00:08:23.668 "rw_ios_per_sec": 0, 00:08:23.668 "rw_mbytes_per_sec": 0, 00:08:23.668 "r_mbytes_per_sec": 0, 00:08:23.668 "w_mbytes_per_sec": 0 00:08:23.668 }, 00:08:23.668 "claimed": false, 00:08:23.668 "zoned": false, 00:08:23.668 "supported_io_types": { 00:08:23.668 "read": true, 00:08:23.668 "write": true, 00:08:23.668 "unmap": true, 00:08:23.668 "flush": true, 00:08:23.668 "reset": true, 00:08:23.668 "nvme_admin": false, 00:08:23.668 "nvme_io": false, 00:08:23.668 "nvme_io_md": false, 00:08:23.668 "write_zeroes": true, 00:08:23.668 "zcopy": false, 00:08:23.668 "get_zone_info": false, 00:08:23.668 "zone_management": false, 00:08:23.668 "zone_append": false, 00:08:23.668 "compare": false, 00:08:23.668 "compare_and_write": false, 00:08:23.668 "abort": false, 00:08:23.668 "seek_hole": false, 00:08:23.668 "seek_data": false, 00:08:23.668 "copy": false, 00:08:23.668 "nvme_iov_md": false 00:08:23.668 }, 00:08:23.668 "memory_domains": [ 00:08:23.668 { 00:08:23.668 "dma_device_id": "system", 00:08:23.668 "dma_device_type": 1 00:08:23.668 }, 00:08:23.668 { 00:08:23.668 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.668 "dma_device_type": 2 00:08:23.668 }, 00:08:23.668 { 00:08:23.668 "dma_device_id": "system", 00:08:23.668 "dma_device_type": 1 00:08:23.668 }, 00:08:23.668 { 00:08:23.668 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.668 "dma_device_type": 2 00:08:23.668 }, 00:08:23.668 { 00:08:23.668 "dma_device_id": "system", 00:08:23.668 "dma_device_type": 1 00:08:23.668 }, 00:08:23.668 { 00:08:23.668 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.668 "dma_device_type": 2 00:08:23.668 } 00:08:23.668 ], 00:08:23.668 "driver_specific": { 00:08:23.668 "raid": { 00:08:23.668 "uuid": "df78b5a1-c9b2-4ca1-9de5-776dce56606f", 00:08:23.668 "strip_size_kb": 64, 00:08:23.668 "state": "online", 00:08:23.668 "raid_level": "concat", 00:08:23.668 "superblock": false, 00:08:23.668 "num_base_bdevs": 3, 00:08:23.668 "num_base_bdevs_discovered": 3, 00:08:23.668 "num_base_bdevs_operational": 3, 00:08:23.668 "base_bdevs_list": [ 00:08:23.668 { 00:08:23.668 "name": "NewBaseBdev", 00:08:23.668 "uuid": "500dce35-dd74-414f-a9c7-73a73b654232", 00:08:23.668 "is_configured": true, 00:08:23.668 "data_offset": 0, 00:08:23.668 "data_size": 65536 00:08:23.668 }, 00:08:23.668 { 00:08:23.668 "name": "BaseBdev2", 00:08:23.668 "uuid": "ebec53cd-45f9-4df2-9fde-7d292a85a0de", 00:08:23.668 "is_configured": true, 00:08:23.668 "data_offset": 0, 00:08:23.668 "data_size": 65536 00:08:23.668 }, 00:08:23.668 { 00:08:23.668 "name": "BaseBdev3", 00:08:23.668 "uuid": "5c2cd370-3a48-4ab2-94f3-cf002f9c9832", 00:08:23.668 "is_configured": true, 00:08:23.668 "data_offset": 0, 00:08:23.668 "data_size": 65536 00:08:23.668 } 00:08:23.668 ] 00:08:23.668 } 00:08:23.668 } 00:08:23.668 }' 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:23.668 BaseBdev2 00:08:23.668 BaseBdev3' 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:23.668 05:35:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:23.668 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:23.668 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:23.668 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.668 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.668 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.668 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.927 [2024-12-07 05:35:57.085152] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:23.927 [2024-12-07 05:35:57.085183] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:23.927 [2024-12-07 05:35:57.085276] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:23.927 [2024-12-07 05:35:57.085333] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:23.927 [2024-12-07 05:35:57.085346] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 76419 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 76419 ']' 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 76419 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76419 00:08:23.927 killing process with pid 76419 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76419' 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 76419 00:08:23.927 [2024-12-07 05:35:57.119858] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:23.927 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 76419 00:08:23.927 [2024-12-07 05:35:57.151284] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:24.185 ************************************ 00:08:24.185 END TEST raid_state_function_test 00:08:24.185 ************************************ 00:08:24.185 05:35:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:24.185 00:08:24.185 real 0m8.360s 00:08:24.185 user 0m14.399s 00:08:24.185 sys 0m1.587s 00:08:24.185 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:24.185 05:35:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.185 05:35:57 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:08:24.185 05:35:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:24.185 05:35:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:24.185 05:35:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:24.185 ************************************ 00:08:24.185 START TEST raid_state_function_test_sb 00:08:24.185 ************************************ 00:08:24.185 05:35:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 true 00:08:24.185 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:24.185 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:24.185 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:24.185 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:24.185 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:24.186 Process raid pid: 77018 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=77018 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 77018' 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 77018 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 77018 ']' 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:24.186 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:24.186 05:35:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.186 [2024-12-07 05:35:57.519117] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:08:24.186 [2024-12-07 05:35:57.519315] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:24.446 [2024-12-07 05:35:57.673390] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:24.446 [2024-12-07 05:35:57.699881] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:24.446 [2024-12-07 05:35:57.742281] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:24.446 [2024-12-07 05:35:57.742403] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:25.014 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:25.014 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:08:25.014 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:25.014 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.014 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.014 [2024-12-07 05:35:58.356943] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:25.014 [2024-12-07 05:35:58.357075] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:25.014 [2024-12-07 05:35:58.357134] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:25.015 [2024-12-07 05:35:58.357159] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:25.015 [2024-12-07 05:35:58.357197] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:25.015 [2024-12-07 05:35:58.357223] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.015 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.274 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.274 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.274 "name": "Existed_Raid", 00:08:25.274 "uuid": "71902990-acf0-4c19-b601-ffc81ce10bb8", 00:08:25.274 "strip_size_kb": 64, 00:08:25.274 "state": "configuring", 00:08:25.274 "raid_level": "concat", 00:08:25.274 "superblock": true, 00:08:25.274 "num_base_bdevs": 3, 00:08:25.274 "num_base_bdevs_discovered": 0, 00:08:25.274 "num_base_bdevs_operational": 3, 00:08:25.274 "base_bdevs_list": [ 00:08:25.274 { 00:08:25.274 "name": "BaseBdev1", 00:08:25.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.274 "is_configured": false, 00:08:25.274 "data_offset": 0, 00:08:25.274 "data_size": 0 00:08:25.274 }, 00:08:25.274 { 00:08:25.274 "name": "BaseBdev2", 00:08:25.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.274 "is_configured": false, 00:08:25.274 "data_offset": 0, 00:08:25.274 "data_size": 0 00:08:25.274 }, 00:08:25.274 { 00:08:25.274 "name": "BaseBdev3", 00:08:25.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.274 "is_configured": false, 00:08:25.274 "data_offset": 0, 00:08:25.274 "data_size": 0 00:08:25.274 } 00:08:25.274 ] 00:08:25.274 }' 00:08:25.274 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.274 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.533 [2024-12-07 05:35:58.716206] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:25.533 [2024-12-07 05:35:58.716250] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.533 [2024-12-07 05:35:58.728214] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:25.533 [2024-12-07 05:35:58.728299] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:25.533 [2024-12-07 05:35:58.728311] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:25.533 [2024-12-07 05:35:58.728336] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:25.533 [2024-12-07 05:35:58.728343] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:25.533 [2024-12-07 05:35:58.728351] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.533 [2024-12-07 05:35:58.749132] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:25.533 BaseBdev1 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.533 [ 00:08:25.533 { 00:08:25.533 "name": "BaseBdev1", 00:08:25.533 "aliases": [ 00:08:25.533 "ea7f609f-a7c2-4474-af6e-af40a126a4f1" 00:08:25.533 ], 00:08:25.533 "product_name": "Malloc disk", 00:08:25.533 "block_size": 512, 00:08:25.533 "num_blocks": 65536, 00:08:25.533 "uuid": "ea7f609f-a7c2-4474-af6e-af40a126a4f1", 00:08:25.533 "assigned_rate_limits": { 00:08:25.533 "rw_ios_per_sec": 0, 00:08:25.533 "rw_mbytes_per_sec": 0, 00:08:25.533 "r_mbytes_per_sec": 0, 00:08:25.533 "w_mbytes_per_sec": 0 00:08:25.533 }, 00:08:25.533 "claimed": true, 00:08:25.533 "claim_type": "exclusive_write", 00:08:25.533 "zoned": false, 00:08:25.533 "supported_io_types": { 00:08:25.533 "read": true, 00:08:25.533 "write": true, 00:08:25.533 "unmap": true, 00:08:25.533 "flush": true, 00:08:25.533 "reset": true, 00:08:25.533 "nvme_admin": false, 00:08:25.533 "nvme_io": false, 00:08:25.533 "nvme_io_md": false, 00:08:25.533 "write_zeroes": true, 00:08:25.533 "zcopy": true, 00:08:25.533 "get_zone_info": false, 00:08:25.533 "zone_management": false, 00:08:25.533 "zone_append": false, 00:08:25.533 "compare": false, 00:08:25.533 "compare_and_write": false, 00:08:25.533 "abort": true, 00:08:25.533 "seek_hole": false, 00:08:25.533 "seek_data": false, 00:08:25.533 "copy": true, 00:08:25.533 "nvme_iov_md": false 00:08:25.533 }, 00:08:25.533 "memory_domains": [ 00:08:25.533 { 00:08:25.533 "dma_device_id": "system", 00:08:25.533 "dma_device_type": 1 00:08:25.533 }, 00:08:25.533 { 00:08:25.533 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:25.533 "dma_device_type": 2 00:08:25.533 } 00:08:25.533 ], 00:08:25.533 "driver_specific": {} 00:08:25.533 } 00:08:25.533 ] 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.533 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.533 "name": "Existed_Raid", 00:08:25.533 "uuid": "10178a07-19ea-44d5-b147-b5d1de17677b", 00:08:25.533 "strip_size_kb": 64, 00:08:25.533 "state": "configuring", 00:08:25.533 "raid_level": "concat", 00:08:25.533 "superblock": true, 00:08:25.533 "num_base_bdevs": 3, 00:08:25.534 "num_base_bdevs_discovered": 1, 00:08:25.534 "num_base_bdevs_operational": 3, 00:08:25.534 "base_bdevs_list": [ 00:08:25.534 { 00:08:25.534 "name": "BaseBdev1", 00:08:25.534 "uuid": "ea7f609f-a7c2-4474-af6e-af40a126a4f1", 00:08:25.534 "is_configured": true, 00:08:25.534 "data_offset": 2048, 00:08:25.534 "data_size": 63488 00:08:25.534 }, 00:08:25.534 { 00:08:25.534 "name": "BaseBdev2", 00:08:25.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.534 "is_configured": false, 00:08:25.534 "data_offset": 0, 00:08:25.534 "data_size": 0 00:08:25.534 }, 00:08:25.534 { 00:08:25.534 "name": "BaseBdev3", 00:08:25.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.534 "is_configured": false, 00:08:25.534 "data_offset": 0, 00:08:25.534 "data_size": 0 00:08:25.534 } 00:08:25.534 ] 00:08:25.534 }' 00:08:25.534 05:35:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.534 05:35:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.123 [2024-12-07 05:35:59.208423] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:26.123 [2024-12-07 05:35:59.208532] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.123 [2024-12-07 05:35:59.216447] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:26.123 [2024-12-07 05:35:59.218327] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:26.123 [2024-12-07 05:35:59.218407] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:26.123 [2024-12-07 05:35:59.218463] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:26.123 [2024-12-07 05:35:59.218489] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.123 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.124 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.124 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.124 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.124 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.124 "name": "Existed_Raid", 00:08:26.124 "uuid": "553e7625-dfd9-41ac-8a73-7edc60d481c7", 00:08:26.124 "strip_size_kb": 64, 00:08:26.124 "state": "configuring", 00:08:26.124 "raid_level": "concat", 00:08:26.124 "superblock": true, 00:08:26.124 "num_base_bdevs": 3, 00:08:26.124 "num_base_bdevs_discovered": 1, 00:08:26.124 "num_base_bdevs_operational": 3, 00:08:26.124 "base_bdevs_list": [ 00:08:26.124 { 00:08:26.124 "name": "BaseBdev1", 00:08:26.124 "uuid": "ea7f609f-a7c2-4474-af6e-af40a126a4f1", 00:08:26.124 "is_configured": true, 00:08:26.124 "data_offset": 2048, 00:08:26.124 "data_size": 63488 00:08:26.124 }, 00:08:26.124 { 00:08:26.124 "name": "BaseBdev2", 00:08:26.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.124 "is_configured": false, 00:08:26.124 "data_offset": 0, 00:08:26.124 "data_size": 0 00:08:26.124 }, 00:08:26.124 { 00:08:26.124 "name": "BaseBdev3", 00:08:26.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.124 "is_configured": false, 00:08:26.124 "data_offset": 0, 00:08:26.124 "data_size": 0 00:08:26.124 } 00:08:26.124 ] 00:08:26.124 }' 00:08:26.124 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.124 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.383 [2024-12-07 05:35:59.678848] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:26.383 BaseBdev2 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.383 [ 00:08:26.383 { 00:08:26.383 "name": "BaseBdev2", 00:08:26.383 "aliases": [ 00:08:26.383 "1afbff1f-6632-478b-8a51-acbb7a356454" 00:08:26.383 ], 00:08:26.383 "product_name": "Malloc disk", 00:08:26.383 "block_size": 512, 00:08:26.383 "num_blocks": 65536, 00:08:26.383 "uuid": "1afbff1f-6632-478b-8a51-acbb7a356454", 00:08:26.383 "assigned_rate_limits": { 00:08:26.383 "rw_ios_per_sec": 0, 00:08:26.383 "rw_mbytes_per_sec": 0, 00:08:26.383 "r_mbytes_per_sec": 0, 00:08:26.383 "w_mbytes_per_sec": 0 00:08:26.383 }, 00:08:26.383 "claimed": true, 00:08:26.383 "claim_type": "exclusive_write", 00:08:26.383 "zoned": false, 00:08:26.383 "supported_io_types": { 00:08:26.383 "read": true, 00:08:26.383 "write": true, 00:08:26.383 "unmap": true, 00:08:26.383 "flush": true, 00:08:26.383 "reset": true, 00:08:26.383 "nvme_admin": false, 00:08:26.383 "nvme_io": false, 00:08:26.383 "nvme_io_md": false, 00:08:26.383 "write_zeroes": true, 00:08:26.383 "zcopy": true, 00:08:26.383 "get_zone_info": false, 00:08:26.383 "zone_management": false, 00:08:26.383 "zone_append": false, 00:08:26.383 "compare": false, 00:08:26.383 "compare_and_write": false, 00:08:26.383 "abort": true, 00:08:26.383 "seek_hole": false, 00:08:26.383 "seek_data": false, 00:08:26.383 "copy": true, 00:08:26.383 "nvme_iov_md": false 00:08:26.383 }, 00:08:26.383 "memory_domains": [ 00:08:26.383 { 00:08:26.383 "dma_device_id": "system", 00:08:26.383 "dma_device_type": 1 00:08:26.383 }, 00:08:26.383 { 00:08:26.383 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:26.383 "dma_device_type": 2 00:08:26.383 } 00:08:26.383 ], 00:08:26.383 "driver_specific": {} 00:08:26.383 } 00:08:26.383 ] 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.383 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.384 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.641 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.641 "name": "Existed_Raid", 00:08:26.641 "uuid": "553e7625-dfd9-41ac-8a73-7edc60d481c7", 00:08:26.641 "strip_size_kb": 64, 00:08:26.641 "state": "configuring", 00:08:26.641 "raid_level": "concat", 00:08:26.641 "superblock": true, 00:08:26.641 "num_base_bdevs": 3, 00:08:26.641 "num_base_bdevs_discovered": 2, 00:08:26.641 "num_base_bdevs_operational": 3, 00:08:26.641 "base_bdevs_list": [ 00:08:26.642 { 00:08:26.642 "name": "BaseBdev1", 00:08:26.642 "uuid": "ea7f609f-a7c2-4474-af6e-af40a126a4f1", 00:08:26.642 "is_configured": true, 00:08:26.642 "data_offset": 2048, 00:08:26.642 "data_size": 63488 00:08:26.642 }, 00:08:26.642 { 00:08:26.642 "name": "BaseBdev2", 00:08:26.642 "uuid": "1afbff1f-6632-478b-8a51-acbb7a356454", 00:08:26.642 "is_configured": true, 00:08:26.642 "data_offset": 2048, 00:08:26.642 "data_size": 63488 00:08:26.642 }, 00:08:26.642 { 00:08:26.642 "name": "BaseBdev3", 00:08:26.642 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.642 "is_configured": false, 00:08:26.642 "data_offset": 0, 00:08:26.642 "data_size": 0 00:08:26.642 } 00:08:26.642 ] 00:08:26.642 }' 00:08:26.642 05:35:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.642 05:35:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.900 [2024-12-07 05:36:00.126686] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:26.900 [2024-12-07 05:36:00.127021] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:26.900 [2024-12-07 05:36:00.127088] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:26.900 [2024-12-07 05:36:00.127449] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:26.900 BaseBdev3 00:08:26.900 [2024-12-07 05:36:00.127670] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:26.900 [2024-12-07 05:36:00.127692] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:26.900 [2024-12-07 05:36:00.127835] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.900 [ 00:08:26.900 { 00:08:26.900 "name": "BaseBdev3", 00:08:26.900 "aliases": [ 00:08:26.900 "ac35ebc2-766a-4860-8be0-46ddaab894ed" 00:08:26.900 ], 00:08:26.900 "product_name": "Malloc disk", 00:08:26.900 "block_size": 512, 00:08:26.900 "num_blocks": 65536, 00:08:26.900 "uuid": "ac35ebc2-766a-4860-8be0-46ddaab894ed", 00:08:26.900 "assigned_rate_limits": { 00:08:26.900 "rw_ios_per_sec": 0, 00:08:26.900 "rw_mbytes_per_sec": 0, 00:08:26.900 "r_mbytes_per_sec": 0, 00:08:26.900 "w_mbytes_per_sec": 0 00:08:26.900 }, 00:08:26.900 "claimed": true, 00:08:26.900 "claim_type": "exclusive_write", 00:08:26.900 "zoned": false, 00:08:26.900 "supported_io_types": { 00:08:26.900 "read": true, 00:08:26.900 "write": true, 00:08:26.900 "unmap": true, 00:08:26.900 "flush": true, 00:08:26.900 "reset": true, 00:08:26.900 "nvme_admin": false, 00:08:26.900 "nvme_io": false, 00:08:26.900 "nvme_io_md": false, 00:08:26.900 "write_zeroes": true, 00:08:26.900 "zcopy": true, 00:08:26.900 "get_zone_info": false, 00:08:26.900 "zone_management": false, 00:08:26.900 "zone_append": false, 00:08:26.900 "compare": false, 00:08:26.900 "compare_and_write": false, 00:08:26.900 "abort": true, 00:08:26.900 "seek_hole": false, 00:08:26.900 "seek_data": false, 00:08:26.900 "copy": true, 00:08:26.900 "nvme_iov_md": false 00:08:26.900 }, 00:08:26.900 "memory_domains": [ 00:08:26.900 { 00:08:26.900 "dma_device_id": "system", 00:08:26.900 "dma_device_type": 1 00:08:26.900 }, 00:08:26.900 { 00:08:26.900 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:26.900 "dma_device_type": 2 00:08:26.900 } 00:08:26.900 ], 00:08:26.900 "driver_specific": {} 00:08:26.900 } 00:08:26.900 ] 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.900 "name": "Existed_Raid", 00:08:26.900 "uuid": "553e7625-dfd9-41ac-8a73-7edc60d481c7", 00:08:26.900 "strip_size_kb": 64, 00:08:26.900 "state": "online", 00:08:26.900 "raid_level": "concat", 00:08:26.900 "superblock": true, 00:08:26.900 "num_base_bdevs": 3, 00:08:26.900 "num_base_bdevs_discovered": 3, 00:08:26.900 "num_base_bdevs_operational": 3, 00:08:26.900 "base_bdevs_list": [ 00:08:26.900 { 00:08:26.900 "name": "BaseBdev1", 00:08:26.900 "uuid": "ea7f609f-a7c2-4474-af6e-af40a126a4f1", 00:08:26.900 "is_configured": true, 00:08:26.900 "data_offset": 2048, 00:08:26.900 "data_size": 63488 00:08:26.900 }, 00:08:26.900 { 00:08:26.900 "name": "BaseBdev2", 00:08:26.900 "uuid": "1afbff1f-6632-478b-8a51-acbb7a356454", 00:08:26.900 "is_configured": true, 00:08:26.900 "data_offset": 2048, 00:08:26.900 "data_size": 63488 00:08:26.900 }, 00:08:26.900 { 00:08:26.900 "name": "BaseBdev3", 00:08:26.900 "uuid": "ac35ebc2-766a-4860-8be0-46ddaab894ed", 00:08:26.900 "is_configured": true, 00:08:26.900 "data_offset": 2048, 00:08:26.900 "data_size": 63488 00:08:26.900 } 00:08:26.900 ] 00:08:26.900 }' 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.900 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:27.467 [2024-12-07 05:36:00.622195] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.467 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:27.467 "name": "Existed_Raid", 00:08:27.467 "aliases": [ 00:08:27.467 "553e7625-dfd9-41ac-8a73-7edc60d481c7" 00:08:27.467 ], 00:08:27.467 "product_name": "Raid Volume", 00:08:27.467 "block_size": 512, 00:08:27.467 "num_blocks": 190464, 00:08:27.467 "uuid": "553e7625-dfd9-41ac-8a73-7edc60d481c7", 00:08:27.467 "assigned_rate_limits": { 00:08:27.467 "rw_ios_per_sec": 0, 00:08:27.467 "rw_mbytes_per_sec": 0, 00:08:27.467 "r_mbytes_per_sec": 0, 00:08:27.467 "w_mbytes_per_sec": 0 00:08:27.467 }, 00:08:27.467 "claimed": false, 00:08:27.467 "zoned": false, 00:08:27.467 "supported_io_types": { 00:08:27.467 "read": true, 00:08:27.467 "write": true, 00:08:27.467 "unmap": true, 00:08:27.467 "flush": true, 00:08:27.467 "reset": true, 00:08:27.467 "nvme_admin": false, 00:08:27.467 "nvme_io": false, 00:08:27.467 "nvme_io_md": false, 00:08:27.467 "write_zeroes": true, 00:08:27.467 "zcopy": false, 00:08:27.467 "get_zone_info": false, 00:08:27.467 "zone_management": false, 00:08:27.467 "zone_append": false, 00:08:27.467 "compare": false, 00:08:27.467 "compare_and_write": false, 00:08:27.467 "abort": false, 00:08:27.467 "seek_hole": false, 00:08:27.467 "seek_data": false, 00:08:27.467 "copy": false, 00:08:27.467 "nvme_iov_md": false 00:08:27.467 }, 00:08:27.467 "memory_domains": [ 00:08:27.467 { 00:08:27.467 "dma_device_id": "system", 00:08:27.467 "dma_device_type": 1 00:08:27.467 }, 00:08:27.467 { 00:08:27.467 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.467 "dma_device_type": 2 00:08:27.467 }, 00:08:27.467 { 00:08:27.467 "dma_device_id": "system", 00:08:27.467 "dma_device_type": 1 00:08:27.467 }, 00:08:27.467 { 00:08:27.467 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.467 "dma_device_type": 2 00:08:27.467 }, 00:08:27.467 { 00:08:27.467 "dma_device_id": "system", 00:08:27.467 "dma_device_type": 1 00:08:27.467 }, 00:08:27.467 { 00:08:27.467 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.467 "dma_device_type": 2 00:08:27.467 } 00:08:27.467 ], 00:08:27.467 "driver_specific": { 00:08:27.467 "raid": { 00:08:27.467 "uuid": "553e7625-dfd9-41ac-8a73-7edc60d481c7", 00:08:27.467 "strip_size_kb": 64, 00:08:27.467 "state": "online", 00:08:27.467 "raid_level": "concat", 00:08:27.467 "superblock": true, 00:08:27.467 "num_base_bdevs": 3, 00:08:27.467 "num_base_bdevs_discovered": 3, 00:08:27.467 "num_base_bdevs_operational": 3, 00:08:27.467 "base_bdevs_list": [ 00:08:27.467 { 00:08:27.467 "name": "BaseBdev1", 00:08:27.467 "uuid": "ea7f609f-a7c2-4474-af6e-af40a126a4f1", 00:08:27.467 "is_configured": true, 00:08:27.467 "data_offset": 2048, 00:08:27.467 "data_size": 63488 00:08:27.467 }, 00:08:27.467 { 00:08:27.467 "name": "BaseBdev2", 00:08:27.467 "uuid": "1afbff1f-6632-478b-8a51-acbb7a356454", 00:08:27.467 "is_configured": true, 00:08:27.467 "data_offset": 2048, 00:08:27.467 "data_size": 63488 00:08:27.467 }, 00:08:27.467 { 00:08:27.467 "name": "BaseBdev3", 00:08:27.468 "uuid": "ac35ebc2-766a-4860-8be0-46ddaab894ed", 00:08:27.468 "is_configured": true, 00:08:27.468 "data_offset": 2048, 00:08:27.468 "data_size": 63488 00:08:27.468 } 00:08:27.468 ] 00:08:27.468 } 00:08:27.468 } 00:08:27.468 }' 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:27.468 BaseBdev2 00:08:27.468 BaseBdev3' 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.468 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.726 [2024-12-07 05:36:00.857534] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:27.726 [2024-12-07 05:36:00.857564] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:27.726 [2024-12-07 05:36:00.857651] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.726 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.727 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:27.727 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.727 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.727 "name": "Existed_Raid", 00:08:27.727 "uuid": "553e7625-dfd9-41ac-8a73-7edc60d481c7", 00:08:27.727 "strip_size_kb": 64, 00:08:27.727 "state": "offline", 00:08:27.727 "raid_level": "concat", 00:08:27.727 "superblock": true, 00:08:27.727 "num_base_bdevs": 3, 00:08:27.727 "num_base_bdevs_discovered": 2, 00:08:27.727 "num_base_bdevs_operational": 2, 00:08:27.727 "base_bdevs_list": [ 00:08:27.727 { 00:08:27.727 "name": null, 00:08:27.727 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.727 "is_configured": false, 00:08:27.727 "data_offset": 0, 00:08:27.727 "data_size": 63488 00:08:27.727 }, 00:08:27.727 { 00:08:27.727 "name": "BaseBdev2", 00:08:27.727 "uuid": "1afbff1f-6632-478b-8a51-acbb7a356454", 00:08:27.727 "is_configured": true, 00:08:27.727 "data_offset": 2048, 00:08:27.727 "data_size": 63488 00:08:27.727 }, 00:08:27.727 { 00:08:27.727 "name": "BaseBdev3", 00:08:27.727 "uuid": "ac35ebc2-766a-4860-8be0-46ddaab894ed", 00:08:27.727 "is_configured": true, 00:08:27.727 "data_offset": 2048, 00:08:27.727 "data_size": 63488 00:08:27.727 } 00:08:27.727 ] 00:08:27.727 }' 00:08:27.727 05:36:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.727 05:36:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.985 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:27.985 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:27.985 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:27.985 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.985 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.985 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.985 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.244 [2024-12-07 05:36:01.376163] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.244 [2024-12-07 05:36:01.435445] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:28.244 [2024-12-07 05:36:01.435494] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.244 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.245 BaseBdev2 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.245 [ 00:08:28.245 { 00:08:28.245 "name": "BaseBdev2", 00:08:28.245 "aliases": [ 00:08:28.245 "9a242cdd-5ab8-4ec1-b3e7-d2849664c29d" 00:08:28.245 ], 00:08:28.245 "product_name": "Malloc disk", 00:08:28.245 "block_size": 512, 00:08:28.245 "num_blocks": 65536, 00:08:28.245 "uuid": "9a242cdd-5ab8-4ec1-b3e7-d2849664c29d", 00:08:28.245 "assigned_rate_limits": { 00:08:28.245 "rw_ios_per_sec": 0, 00:08:28.245 "rw_mbytes_per_sec": 0, 00:08:28.245 "r_mbytes_per_sec": 0, 00:08:28.245 "w_mbytes_per_sec": 0 00:08:28.245 }, 00:08:28.245 "claimed": false, 00:08:28.245 "zoned": false, 00:08:28.245 "supported_io_types": { 00:08:28.245 "read": true, 00:08:28.245 "write": true, 00:08:28.245 "unmap": true, 00:08:28.245 "flush": true, 00:08:28.245 "reset": true, 00:08:28.245 "nvme_admin": false, 00:08:28.245 "nvme_io": false, 00:08:28.245 "nvme_io_md": false, 00:08:28.245 "write_zeroes": true, 00:08:28.245 "zcopy": true, 00:08:28.245 "get_zone_info": false, 00:08:28.245 "zone_management": false, 00:08:28.245 "zone_append": false, 00:08:28.245 "compare": false, 00:08:28.245 "compare_and_write": false, 00:08:28.245 "abort": true, 00:08:28.245 "seek_hole": false, 00:08:28.245 "seek_data": false, 00:08:28.245 "copy": true, 00:08:28.245 "nvme_iov_md": false 00:08:28.245 }, 00:08:28.245 "memory_domains": [ 00:08:28.245 { 00:08:28.245 "dma_device_id": "system", 00:08:28.245 "dma_device_type": 1 00:08:28.245 }, 00:08:28.245 { 00:08:28.245 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.245 "dma_device_type": 2 00:08:28.245 } 00:08:28.245 ], 00:08:28.245 "driver_specific": {} 00:08:28.245 } 00:08:28.245 ] 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.245 BaseBdev3 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.245 [ 00:08:28.245 { 00:08:28.245 "name": "BaseBdev3", 00:08:28.245 "aliases": [ 00:08:28.245 "576d1a6a-177a-442f-b924-e8d3777cf216" 00:08:28.245 ], 00:08:28.245 "product_name": "Malloc disk", 00:08:28.245 "block_size": 512, 00:08:28.245 "num_blocks": 65536, 00:08:28.245 "uuid": "576d1a6a-177a-442f-b924-e8d3777cf216", 00:08:28.245 "assigned_rate_limits": { 00:08:28.245 "rw_ios_per_sec": 0, 00:08:28.245 "rw_mbytes_per_sec": 0, 00:08:28.245 "r_mbytes_per_sec": 0, 00:08:28.245 "w_mbytes_per_sec": 0 00:08:28.245 }, 00:08:28.245 "claimed": false, 00:08:28.245 "zoned": false, 00:08:28.245 "supported_io_types": { 00:08:28.245 "read": true, 00:08:28.245 "write": true, 00:08:28.245 "unmap": true, 00:08:28.245 "flush": true, 00:08:28.245 "reset": true, 00:08:28.245 "nvme_admin": false, 00:08:28.245 "nvme_io": false, 00:08:28.245 "nvme_io_md": false, 00:08:28.245 "write_zeroes": true, 00:08:28.245 "zcopy": true, 00:08:28.245 "get_zone_info": false, 00:08:28.245 "zone_management": false, 00:08:28.245 "zone_append": false, 00:08:28.245 "compare": false, 00:08:28.245 "compare_and_write": false, 00:08:28.245 "abort": true, 00:08:28.245 "seek_hole": false, 00:08:28.245 "seek_data": false, 00:08:28.245 "copy": true, 00:08:28.245 "nvme_iov_md": false 00:08:28.245 }, 00:08:28.245 "memory_domains": [ 00:08:28.245 { 00:08:28.245 "dma_device_id": "system", 00:08:28.245 "dma_device_type": 1 00:08:28.245 }, 00:08:28.245 { 00:08:28.245 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.245 "dma_device_type": 2 00:08:28.245 } 00:08:28.245 ], 00:08:28.245 "driver_specific": {} 00:08:28.245 } 00:08:28.245 ] 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.245 [2024-12-07 05:36:01.600204] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:28.245 [2024-12-07 05:36:01.600290] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:28.245 [2024-12-07 05:36:01.600330] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:28.245 [2024-12-07 05:36:01.602353] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.245 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.503 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.503 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:28.503 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.503 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.503 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.503 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.503 "name": "Existed_Raid", 00:08:28.503 "uuid": "e8c60020-d6ed-44b2-8fa5-6d62e0067783", 00:08:28.503 "strip_size_kb": 64, 00:08:28.503 "state": "configuring", 00:08:28.503 "raid_level": "concat", 00:08:28.503 "superblock": true, 00:08:28.503 "num_base_bdevs": 3, 00:08:28.503 "num_base_bdevs_discovered": 2, 00:08:28.503 "num_base_bdevs_operational": 3, 00:08:28.503 "base_bdevs_list": [ 00:08:28.503 { 00:08:28.503 "name": "BaseBdev1", 00:08:28.503 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:28.503 "is_configured": false, 00:08:28.503 "data_offset": 0, 00:08:28.503 "data_size": 0 00:08:28.503 }, 00:08:28.503 { 00:08:28.503 "name": "BaseBdev2", 00:08:28.503 "uuid": "9a242cdd-5ab8-4ec1-b3e7-d2849664c29d", 00:08:28.503 "is_configured": true, 00:08:28.503 "data_offset": 2048, 00:08:28.503 "data_size": 63488 00:08:28.503 }, 00:08:28.503 { 00:08:28.503 "name": "BaseBdev3", 00:08:28.503 "uuid": "576d1a6a-177a-442f-b924-e8d3777cf216", 00:08:28.503 "is_configured": true, 00:08:28.503 "data_offset": 2048, 00:08:28.503 "data_size": 63488 00:08:28.503 } 00:08:28.503 ] 00:08:28.503 }' 00:08:28.503 05:36:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:28.503 05:36:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.761 [2024-12-07 05:36:02.031504] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.761 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.761 "name": "Existed_Raid", 00:08:28.761 "uuid": "e8c60020-d6ed-44b2-8fa5-6d62e0067783", 00:08:28.761 "strip_size_kb": 64, 00:08:28.761 "state": "configuring", 00:08:28.761 "raid_level": "concat", 00:08:28.761 "superblock": true, 00:08:28.761 "num_base_bdevs": 3, 00:08:28.761 "num_base_bdevs_discovered": 1, 00:08:28.761 "num_base_bdevs_operational": 3, 00:08:28.761 "base_bdevs_list": [ 00:08:28.761 { 00:08:28.761 "name": "BaseBdev1", 00:08:28.761 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:28.761 "is_configured": false, 00:08:28.761 "data_offset": 0, 00:08:28.761 "data_size": 0 00:08:28.761 }, 00:08:28.761 { 00:08:28.761 "name": null, 00:08:28.761 "uuid": "9a242cdd-5ab8-4ec1-b3e7-d2849664c29d", 00:08:28.761 "is_configured": false, 00:08:28.761 "data_offset": 0, 00:08:28.761 "data_size": 63488 00:08:28.761 }, 00:08:28.761 { 00:08:28.762 "name": "BaseBdev3", 00:08:28.762 "uuid": "576d1a6a-177a-442f-b924-e8d3777cf216", 00:08:28.762 "is_configured": true, 00:08:28.762 "data_offset": 2048, 00:08:28.762 "data_size": 63488 00:08:28.762 } 00:08:28.762 ] 00:08:28.762 }' 00:08:28.762 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:28.762 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.328 [2024-12-07 05:36:02.481676] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:29.328 BaseBdev1 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.328 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.328 [ 00:08:29.328 { 00:08:29.328 "name": "BaseBdev1", 00:08:29.328 "aliases": [ 00:08:29.328 "47381092-4e86-4794-b21d-a38baf0b98c2" 00:08:29.328 ], 00:08:29.328 "product_name": "Malloc disk", 00:08:29.328 "block_size": 512, 00:08:29.328 "num_blocks": 65536, 00:08:29.328 "uuid": "47381092-4e86-4794-b21d-a38baf0b98c2", 00:08:29.328 "assigned_rate_limits": { 00:08:29.328 "rw_ios_per_sec": 0, 00:08:29.328 "rw_mbytes_per_sec": 0, 00:08:29.328 "r_mbytes_per_sec": 0, 00:08:29.328 "w_mbytes_per_sec": 0 00:08:29.328 }, 00:08:29.328 "claimed": true, 00:08:29.328 "claim_type": "exclusive_write", 00:08:29.328 "zoned": false, 00:08:29.328 "supported_io_types": { 00:08:29.328 "read": true, 00:08:29.328 "write": true, 00:08:29.328 "unmap": true, 00:08:29.328 "flush": true, 00:08:29.328 "reset": true, 00:08:29.328 "nvme_admin": false, 00:08:29.328 "nvme_io": false, 00:08:29.328 "nvme_io_md": false, 00:08:29.328 "write_zeroes": true, 00:08:29.328 "zcopy": true, 00:08:29.328 "get_zone_info": false, 00:08:29.328 "zone_management": false, 00:08:29.328 "zone_append": false, 00:08:29.329 "compare": false, 00:08:29.329 "compare_and_write": false, 00:08:29.329 "abort": true, 00:08:29.329 "seek_hole": false, 00:08:29.329 "seek_data": false, 00:08:29.329 "copy": true, 00:08:29.329 "nvme_iov_md": false 00:08:29.329 }, 00:08:29.329 "memory_domains": [ 00:08:29.329 { 00:08:29.329 "dma_device_id": "system", 00:08:29.329 "dma_device_type": 1 00:08:29.329 }, 00:08:29.329 { 00:08:29.329 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.329 "dma_device_type": 2 00:08:29.329 } 00:08:29.329 ], 00:08:29.329 "driver_specific": {} 00:08:29.329 } 00:08:29.329 ] 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:29.329 "name": "Existed_Raid", 00:08:29.329 "uuid": "e8c60020-d6ed-44b2-8fa5-6d62e0067783", 00:08:29.329 "strip_size_kb": 64, 00:08:29.329 "state": "configuring", 00:08:29.329 "raid_level": "concat", 00:08:29.329 "superblock": true, 00:08:29.329 "num_base_bdevs": 3, 00:08:29.329 "num_base_bdevs_discovered": 2, 00:08:29.329 "num_base_bdevs_operational": 3, 00:08:29.329 "base_bdevs_list": [ 00:08:29.329 { 00:08:29.329 "name": "BaseBdev1", 00:08:29.329 "uuid": "47381092-4e86-4794-b21d-a38baf0b98c2", 00:08:29.329 "is_configured": true, 00:08:29.329 "data_offset": 2048, 00:08:29.329 "data_size": 63488 00:08:29.329 }, 00:08:29.329 { 00:08:29.329 "name": null, 00:08:29.329 "uuid": "9a242cdd-5ab8-4ec1-b3e7-d2849664c29d", 00:08:29.329 "is_configured": false, 00:08:29.329 "data_offset": 0, 00:08:29.329 "data_size": 63488 00:08:29.329 }, 00:08:29.329 { 00:08:29.329 "name": "BaseBdev3", 00:08:29.329 "uuid": "576d1a6a-177a-442f-b924-e8d3777cf216", 00:08:29.329 "is_configured": true, 00:08:29.329 "data_offset": 2048, 00:08:29.329 "data_size": 63488 00:08:29.329 } 00:08:29.329 ] 00:08:29.329 }' 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:29.329 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.589 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.589 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.589 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.589 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:29.589 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.589 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:29.589 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:29.589 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.589 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.589 [2024-12-07 05:36:02.952916] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:29.848 "name": "Existed_Raid", 00:08:29.848 "uuid": "e8c60020-d6ed-44b2-8fa5-6d62e0067783", 00:08:29.848 "strip_size_kb": 64, 00:08:29.848 "state": "configuring", 00:08:29.848 "raid_level": "concat", 00:08:29.848 "superblock": true, 00:08:29.848 "num_base_bdevs": 3, 00:08:29.848 "num_base_bdevs_discovered": 1, 00:08:29.848 "num_base_bdevs_operational": 3, 00:08:29.848 "base_bdevs_list": [ 00:08:29.848 { 00:08:29.848 "name": "BaseBdev1", 00:08:29.848 "uuid": "47381092-4e86-4794-b21d-a38baf0b98c2", 00:08:29.848 "is_configured": true, 00:08:29.848 "data_offset": 2048, 00:08:29.848 "data_size": 63488 00:08:29.848 }, 00:08:29.848 { 00:08:29.848 "name": null, 00:08:29.848 "uuid": "9a242cdd-5ab8-4ec1-b3e7-d2849664c29d", 00:08:29.848 "is_configured": false, 00:08:29.848 "data_offset": 0, 00:08:29.848 "data_size": 63488 00:08:29.848 }, 00:08:29.848 { 00:08:29.848 "name": null, 00:08:29.848 "uuid": "576d1a6a-177a-442f-b924-e8d3777cf216", 00:08:29.848 "is_configured": false, 00:08:29.848 "data_offset": 0, 00:08:29.848 "data_size": 63488 00:08:29.848 } 00:08:29.848 ] 00:08:29.848 }' 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:29.848 05:36:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.108 [2024-12-07 05:36:03.428112] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.108 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.367 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.367 "name": "Existed_Raid", 00:08:30.367 "uuid": "e8c60020-d6ed-44b2-8fa5-6d62e0067783", 00:08:30.367 "strip_size_kb": 64, 00:08:30.367 "state": "configuring", 00:08:30.367 "raid_level": "concat", 00:08:30.367 "superblock": true, 00:08:30.367 "num_base_bdevs": 3, 00:08:30.367 "num_base_bdevs_discovered": 2, 00:08:30.367 "num_base_bdevs_operational": 3, 00:08:30.367 "base_bdevs_list": [ 00:08:30.367 { 00:08:30.367 "name": "BaseBdev1", 00:08:30.367 "uuid": "47381092-4e86-4794-b21d-a38baf0b98c2", 00:08:30.367 "is_configured": true, 00:08:30.367 "data_offset": 2048, 00:08:30.367 "data_size": 63488 00:08:30.367 }, 00:08:30.367 { 00:08:30.367 "name": null, 00:08:30.367 "uuid": "9a242cdd-5ab8-4ec1-b3e7-d2849664c29d", 00:08:30.367 "is_configured": false, 00:08:30.367 "data_offset": 0, 00:08:30.367 "data_size": 63488 00:08:30.367 }, 00:08:30.367 { 00:08:30.367 "name": "BaseBdev3", 00:08:30.367 "uuid": "576d1a6a-177a-442f-b924-e8d3777cf216", 00:08:30.367 "is_configured": true, 00:08:30.367 "data_offset": 2048, 00:08:30.367 "data_size": 63488 00:08:30.367 } 00:08:30.367 ] 00:08:30.367 }' 00:08:30.367 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.367 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.628 [2024-12-07 05:36:03.871413] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.628 "name": "Existed_Raid", 00:08:30.628 "uuid": "e8c60020-d6ed-44b2-8fa5-6d62e0067783", 00:08:30.628 "strip_size_kb": 64, 00:08:30.628 "state": "configuring", 00:08:30.628 "raid_level": "concat", 00:08:30.628 "superblock": true, 00:08:30.628 "num_base_bdevs": 3, 00:08:30.628 "num_base_bdevs_discovered": 1, 00:08:30.628 "num_base_bdevs_operational": 3, 00:08:30.628 "base_bdevs_list": [ 00:08:30.628 { 00:08:30.628 "name": null, 00:08:30.628 "uuid": "47381092-4e86-4794-b21d-a38baf0b98c2", 00:08:30.628 "is_configured": false, 00:08:30.628 "data_offset": 0, 00:08:30.628 "data_size": 63488 00:08:30.628 }, 00:08:30.628 { 00:08:30.628 "name": null, 00:08:30.628 "uuid": "9a242cdd-5ab8-4ec1-b3e7-d2849664c29d", 00:08:30.628 "is_configured": false, 00:08:30.628 "data_offset": 0, 00:08:30.628 "data_size": 63488 00:08:30.628 }, 00:08:30.628 { 00:08:30.628 "name": "BaseBdev3", 00:08:30.628 "uuid": "576d1a6a-177a-442f-b924-e8d3777cf216", 00:08:30.628 "is_configured": true, 00:08:30.628 "data_offset": 2048, 00:08:30.628 "data_size": 63488 00:08:30.628 } 00:08:30.628 ] 00:08:30.628 }' 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.628 05:36:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.197 [2024-12-07 05:36:04.324935] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.197 "name": "Existed_Raid", 00:08:31.197 "uuid": "e8c60020-d6ed-44b2-8fa5-6d62e0067783", 00:08:31.197 "strip_size_kb": 64, 00:08:31.197 "state": "configuring", 00:08:31.197 "raid_level": "concat", 00:08:31.197 "superblock": true, 00:08:31.197 "num_base_bdevs": 3, 00:08:31.197 "num_base_bdevs_discovered": 2, 00:08:31.197 "num_base_bdevs_operational": 3, 00:08:31.197 "base_bdevs_list": [ 00:08:31.197 { 00:08:31.197 "name": null, 00:08:31.197 "uuid": "47381092-4e86-4794-b21d-a38baf0b98c2", 00:08:31.197 "is_configured": false, 00:08:31.197 "data_offset": 0, 00:08:31.197 "data_size": 63488 00:08:31.197 }, 00:08:31.197 { 00:08:31.197 "name": "BaseBdev2", 00:08:31.197 "uuid": "9a242cdd-5ab8-4ec1-b3e7-d2849664c29d", 00:08:31.197 "is_configured": true, 00:08:31.197 "data_offset": 2048, 00:08:31.197 "data_size": 63488 00:08:31.197 }, 00:08:31.197 { 00:08:31.197 "name": "BaseBdev3", 00:08:31.197 "uuid": "576d1a6a-177a-442f-b924-e8d3777cf216", 00:08:31.197 "is_configured": true, 00:08:31.197 "data_offset": 2048, 00:08:31.197 "data_size": 63488 00:08:31.197 } 00:08:31.197 ] 00:08:31.197 }' 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.197 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.457 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:31.457 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.457 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.457 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.457 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 47381092-4e86-4794-b21d-a38baf0b98c2 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.716 [2024-12-07 05:36:04.882883] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:31.716 [2024-12-07 05:36:04.883137] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:31.716 [2024-12-07 05:36:04.883191] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:31.716 [2024-12-07 05:36:04.883450] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:31.716 NewBaseBdev 00:08:31.716 [2024-12-07 05:36:04.883606] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:31.716 [2024-12-07 05:36:04.883618] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:31.716 [2024-12-07 05:36:04.883738] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.716 [ 00:08:31.716 { 00:08:31.716 "name": "NewBaseBdev", 00:08:31.716 "aliases": [ 00:08:31.716 "47381092-4e86-4794-b21d-a38baf0b98c2" 00:08:31.716 ], 00:08:31.716 "product_name": "Malloc disk", 00:08:31.716 "block_size": 512, 00:08:31.716 "num_blocks": 65536, 00:08:31.716 "uuid": "47381092-4e86-4794-b21d-a38baf0b98c2", 00:08:31.716 "assigned_rate_limits": { 00:08:31.716 "rw_ios_per_sec": 0, 00:08:31.716 "rw_mbytes_per_sec": 0, 00:08:31.716 "r_mbytes_per_sec": 0, 00:08:31.716 "w_mbytes_per_sec": 0 00:08:31.716 }, 00:08:31.716 "claimed": true, 00:08:31.716 "claim_type": "exclusive_write", 00:08:31.716 "zoned": false, 00:08:31.716 "supported_io_types": { 00:08:31.716 "read": true, 00:08:31.716 "write": true, 00:08:31.716 "unmap": true, 00:08:31.716 "flush": true, 00:08:31.716 "reset": true, 00:08:31.716 "nvme_admin": false, 00:08:31.716 "nvme_io": false, 00:08:31.716 "nvme_io_md": false, 00:08:31.716 "write_zeroes": true, 00:08:31.716 "zcopy": true, 00:08:31.716 "get_zone_info": false, 00:08:31.716 "zone_management": false, 00:08:31.716 "zone_append": false, 00:08:31.716 "compare": false, 00:08:31.716 "compare_and_write": false, 00:08:31.716 "abort": true, 00:08:31.716 "seek_hole": false, 00:08:31.716 "seek_data": false, 00:08:31.716 "copy": true, 00:08:31.716 "nvme_iov_md": false 00:08:31.716 }, 00:08:31.716 "memory_domains": [ 00:08:31.716 { 00:08:31.716 "dma_device_id": "system", 00:08:31.716 "dma_device_type": 1 00:08:31.716 }, 00:08:31.716 { 00:08:31.716 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:31.716 "dma_device_type": 2 00:08:31.716 } 00:08:31.716 ], 00:08:31.716 "driver_specific": {} 00:08:31.716 } 00:08:31.716 ] 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.716 "name": "Existed_Raid", 00:08:31.716 "uuid": "e8c60020-d6ed-44b2-8fa5-6d62e0067783", 00:08:31.716 "strip_size_kb": 64, 00:08:31.716 "state": "online", 00:08:31.716 "raid_level": "concat", 00:08:31.716 "superblock": true, 00:08:31.716 "num_base_bdevs": 3, 00:08:31.716 "num_base_bdevs_discovered": 3, 00:08:31.716 "num_base_bdevs_operational": 3, 00:08:31.716 "base_bdevs_list": [ 00:08:31.716 { 00:08:31.716 "name": "NewBaseBdev", 00:08:31.716 "uuid": "47381092-4e86-4794-b21d-a38baf0b98c2", 00:08:31.716 "is_configured": true, 00:08:31.716 "data_offset": 2048, 00:08:31.716 "data_size": 63488 00:08:31.716 }, 00:08:31.716 { 00:08:31.716 "name": "BaseBdev2", 00:08:31.716 "uuid": "9a242cdd-5ab8-4ec1-b3e7-d2849664c29d", 00:08:31.716 "is_configured": true, 00:08:31.716 "data_offset": 2048, 00:08:31.716 "data_size": 63488 00:08:31.716 }, 00:08:31.716 { 00:08:31.716 "name": "BaseBdev3", 00:08:31.716 "uuid": "576d1a6a-177a-442f-b924-e8d3777cf216", 00:08:31.716 "is_configured": true, 00:08:31.716 "data_offset": 2048, 00:08:31.716 "data_size": 63488 00:08:31.716 } 00:08:31.716 ] 00:08:31.716 }' 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.716 05:36:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.982 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:31.982 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:31.982 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:31.982 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:31.982 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:31.982 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:31.982 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:32.240 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:32.240 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.240 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.240 [2024-12-07 05:36:05.354504] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:32.240 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.240 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:32.241 "name": "Existed_Raid", 00:08:32.241 "aliases": [ 00:08:32.241 "e8c60020-d6ed-44b2-8fa5-6d62e0067783" 00:08:32.241 ], 00:08:32.241 "product_name": "Raid Volume", 00:08:32.241 "block_size": 512, 00:08:32.241 "num_blocks": 190464, 00:08:32.241 "uuid": "e8c60020-d6ed-44b2-8fa5-6d62e0067783", 00:08:32.241 "assigned_rate_limits": { 00:08:32.241 "rw_ios_per_sec": 0, 00:08:32.241 "rw_mbytes_per_sec": 0, 00:08:32.241 "r_mbytes_per_sec": 0, 00:08:32.241 "w_mbytes_per_sec": 0 00:08:32.241 }, 00:08:32.241 "claimed": false, 00:08:32.241 "zoned": false, 00:08:32.241 "supported_io_types": { 00:08:32.241 "read": true, 00:08:32.241 "write": true, 00:08:32.241 "unmap": true, 00:08:32.241 "flush": true, 00:08:32.241 "reset": true, 00:08:32.241 "nvme_admin": false, 00:08:32.241 "nvme_io": false, 00:08:32.241 "nvme_io_md": false, 00:08:32.241 "write_zeroes": true, 00:08:32.241 "zcopy": false, 00:08:32.241 "get_zone_info": false, 00:08:32.241 "zone_management": false, 00:08:32.241 "zone_append": false, 00:08:32.241 "compare": false, 00:08:32.241 "compare_and_write": false, 00:08:32.241 "abort": false, 00:08:32.241 "seek_hole": false, 00:08:32.241 "seek_data": false, 00:08:32.241 "copy": false, 00:08:32.241 "nvme_iov_md": false 00:08:32.241 }, 00:08:32.241 "memory_domains": [ 00:08:32.241 { 00:08:32.241 "dma_device_id": "system", 00:08:32.241 "dma_device_type": 1 00:08:32.241 }, 00:08:32.241 { 00:08:32.241 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.241 "dma_device_type": 2 00:08:32.241 }, 00:08:32.241 { 00:08:32.241 "dma_device_id": "system", 00:08:32.241 "dma_device_type": 1 00:08:32.241 }, 00:08:32.241 { 00:08:32.241 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.241 "dma_device_type": 2 00:08:32.241 }, 00:08:32.241 { 00:08:32.241 "dma_device_id": "system", 00:08:32.241 "dma_device_type": 1 00:08:32.241 }, 00:08:32.241 { 00:08:32.241 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.241 "dma_device_type": 2 00:08:32.241 } 00:08:32.241 ], 00:08:32.241 "driver_specific": { 00:08:32.241 "raid": { 00:08:32.241 "uuid": "e8c60020-d6ed-44b2-8fa5-6d62e0067783", 00:08:32.241 "strip_size_kb": 64, 00:08:32.241 "state": "online", 00:08:32.241 "raid_level": "concat", 00:08:32.241 "superblock": true, 00:08:32.241 "num_base_bdevs": 3, 00:08:32.241 "num_base_bdevs_discovered": 3, 00:08:32.241 "num_base_bdevs_operational": 3, 00:08:32.241 "base_bdevs_list": [ 00:08:32.241 { 00:08:32.241 "name": "NewBaseBdev", 00:08:32.241 "uuid": "47381092-4e86-4794-b21d-a38baf0b98c2", 00:08:32.241 "is_configured": true, 00:08:32.241 "data_offset": 2048, 00:08:32.241 "data_size": 63488 00:08:32.241 }, 00:08:32.241 { 00:08:32.241 "name": "BaseBdev2", 00:08:32.241 "uuid": "9a242cdd-5ab8-4ec1-b3e7-d2849664c29d", 00:08:32.241 "is_configured": true, 00:08:32.241 "data_offset": 2048, 00:08:32.241 "data_size": 63488 00:08:32.241 }, 00:08:32.241 { 00:08:32.241 "name": "BaseBdev3", 00:08:32.241 "uuid": "576d1a6a-177a-442f-b924-e8d3777cf216", 00:08:32.241 "is_configured": true, 00:08:32.241 "data_offset": 2048, 00:08:32.241 "data_size": 63488 00:08:32.241 } 00:08:32.241 ] 00:08:32.241 } 00:08:32.241 } 00:08:32.241 }' 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:32.241 BaseBdev2 00:08:32.241 BaseBdev3' 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.241 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.241 [2024-12-07 05:36:05.605770] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:32.241 [2024-12-07 05:36:05.605846] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:32.241 [2024-12-07 05:36:05.605952] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:32.241 [2024-12-07 05:36:05.606032] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:32.241 [2024-12-07 05:36:05.606079] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 77018 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 77018 ']' 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 77018 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77018 00:08:32.500 killing process with pid 77018 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77018' 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 77018 00:08:32.500 [2024-12-07 05:36:05.651261] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:32.500 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 77018 00:08:32.500 [2024-12-07 05:36:05.682958] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:32.759 05:36:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:32.759 00:08:32.759 real 0m8.471s 00:08:32.759 user 0m14.482s 00:08:32.759 sys 0m1.629s 00:08:32.759 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:32.759 ************************************ 00:08:32.759 END TEST raid_state_function_test_sb 00:08:32.759 ************************************ 00:08:32.759 05:36:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.759 05:36:05 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:08:32.759 05:36:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:08:32.759 05:36:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:32.759 05:36:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:32.759 ************************************ 00:08:32.759 START TEST raid_superblock_test 00:08:32.759 ************************************ 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 3 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=77616 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 77616 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 77616 ']' 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:32.759 05:36:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:32.759 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:32.760 05:36:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:32.760 05:36:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.760 [2024-12-07 05:36:06.040361] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:08:32.760 [2024-12-07 05:36:06.040561] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77616 ] 00:08:33.018 [2024-12-07 05:36:06.195569] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:33.019 [2024-12-07 05:36:06.222528] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:33.019 [2024-12-07 05:36:06.264962] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:33.019 [2024-12-07 05:36:06.264997] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:33.587 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:33.587 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:33.587 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.588 malloc1 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.588 [2024-12-07 05:36:06.896308] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:33.588 [2024-12-07 05:36:06.896414] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.588 [2024-12-07 05:36:06.896451] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:33.588 [2024-12-07 05:36:06.896488] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.588 [2024-12-07 05:36:06.898650] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.588 [2024-12-07 05:36:06.898723] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:33.588 pt1 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.588 malloc2 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.588 [2024-12-07 05:36:06.928743] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:33.588 [2024-12-07 05:36:06.928800] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.588 [2024-12-07 05:36:06.928818] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:33.588 [2024-12-07 05:36:06.928828] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.588 [2024-12-07 05:36:06.930935] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.588 [2024-12-07 05:36:06.931010] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:33.588 pt2 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.588 malloc3 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.588 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.848 [2024-12-07 05:36:06.957183] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:33.848 [2024-12-07 05:36:06.957277] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.848 [2024-12-07 05:36:06.957312] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:33.848 [2024-12-07 05:36:06.957345] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.848 [2024-12-07 05:36:06.959448] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.848 [2024-12-07 05:36:06.959522] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:33.848 pt3 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.848 [2024-12-07 05:36:06.969236] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:33.848 [2024-12-07 05:36:06.971150] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:33.848 [2024-12-07 05:36:06.971243] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:33.848 [2024-12-07 05:36:06.971424] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:33.848 [2024-12-07 05:36:06.971474] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:33.848 [2024-12-07 05:36:06.971779] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:33.848 [2024-12-07 05:36:06.971942] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:33.848 [2024-12-07 05:36:06.971992] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:33.848 [2024-12-07 05:36:06.972137] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.848 05:36:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.848 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:33.848 "name": "raid_bdev1", 00:08:33.848 "uuid": "fb97627a-712b-4bd2-833c-7ed0811a12e1", 00:08:33.848 "strip_size_kb": 64, 00:08:33.848 "state": "online", 00:08:33.848 "raid_level": "concat", 00:08:33.848 "superblock": true, 00:08:33.848 "num_base_bdevs": 3, 00:08:33.848 "num_base_bdevs_discovered": 3, 00:08:33.848 "num_base_bdevs_operational": 3, 00:08:33.848 "base_bdevs_list": [ 00:08:33.848 { 00:08:33.848 "name": "pt1", 00:08:33.848 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:33.848 "is_configured": true, 00:08:33.848 "data_offset": 2048, 00:08:33.848 "data_size": 63488 00:08:33.848 }, 00:08:33.848 { 00:08:33.848 "name": "pt2", 00:08:33.848 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:33.848 "is_configured": true, 00:08:33.848 "data_offset": 2048, 00:08:33.848 "data_size": 63488 00:08:33.848 }, 00:08:33.848 { 00:08:33.848 "name": "pt3", 00:08:33.848 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:33.849 "is_configured": true, 00:08:33.849 "data_offset": 2048, 00:08:33.849 "data_size": 63488 00:08:33.849 } 00:08:33.849 ] 00:08:33.849 }' 00:08:33.849 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:33.849 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.108 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:34.108 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:34.108 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:34.108 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:34.108 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:34.108 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:34.108 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:34.108 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:34.108 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.108 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.108 [2024-12-07 05:36:07.456708] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:34.108 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.368 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:34.368 "name": "raid_bdev1", 00:08:34.368 "aliases": [ 00:08:34.368 "fb97627a-712b-4bd2-833c-7ed0811a12e1" 00:08:34.368 ], 00:08:34.368 "product_name": "Raid Volume", 00:08:34.368 "block_size": 512, 00:08:34.368 "num_blocks": 190464, 00:08:34.368 "uuid": "fb97627a-712b-4bd2-833c-7ed0811a12e1", 00:08:34.368 "assigned_rate_limits": { 00:08:34.368 "rw_ios_per_sec": 0, 00:08:34.368 "rw_mbytes_per_sec": 0, 00:08:34.368 "r_mbytes_per_sec": 0, 00:08:34.368 "w_mbytes_per_sec": 0 00:08:34.368 }, 00:08:34.368 "claimed": false, 00:08:34.368 "zoned": false, 00:08:34.368 "supported_io_types": { 00:08:34.368 "read": true, 00:08:34.368 "write": true, 00:08:34.368 "unmap": true, 00:08:34.368 "flush": true, 00:08:34.368 "reset": true, 00:08:34.368 "nvme_admin": false, 00:08:34.368 "nvme_io": false, 00:08:34.368 "nvme_io_md": false, 00:08:34.368 "write_zeroes": true, 00:08:34.368 "zcopy": false, 00:08:34.368 "get_zone_info": false, 00:08:34.368 "zone_management": false, 00:08:34.368 "zone_append": false, 00:08:34.368 "compare": false, 00:08:34.368 "compare_and_write": false, 00:08:34.368 "abort": false, 00:08:34.368 "seek_hole": false, 00:08:34.368 "seek_data": false, 00:08:34.368 "copy": false, 00:08:34.368 "nvme_iov_md": false 00:08:34.368 }, 00:08:34.368 "memory_domains": [ 00:08:34.368 { 00:08:34.368 "dma_device_id": "system", 00:08:34.368 "dma_device_type": 1 00:08:34.368 }, 00:08:34.368 { 00:08:34.368 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.368 "dma_device_type": 2 00:08:34.368 }, 00:08:34.368 { 00:08:34.368 "dma_device_id": "system", 00:08:34.368 "dma_device_type": 1 00:08:34.368 }, 00:08:34.368 { 00:08:34.368 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.368 "dma_device_type": 2 00:08:34.368 }, 00:08:34.368 { 00:08:34.368 "dma_device_id": "system", 00:08:34.368 "dma_device_type": 1 00:08:34.368 }, 00:08:34.368 { 00:08:34.368 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.368 "dma_device_type": 2 00:08:34.368 } 00:08:34.368 ], 00:08:34.368 "driver_specific": { 00:08:34.368 "raid": { 00:08:34.368 "uuid": "fb97627a-712b-4bd2-833c-7ed0811a12e1", 00:08:34.368 "strip_size_kb": 64, 00:08:34.368 "state": "online", 00:08:34.368 "raid_level": "concat", 00:08:34.368 "superblock": true, 00:08:34.368 "num_base_bdevs": 3, 00:08:34.368 "num_base_bdevs_discovered": 3, 00:08:34.368 "num_base_bdevs_operational": 3, 00:08:34.368 "base_bdevs_list": [ 00:08:34.368 { 00:08:34.368 "name": "pt1", 00:08:34.368 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:34.368 "is_configured": true, 00:08:34.368 "data_offset": 2048, 00:08:34.368 "data_size": 63488 00:08:34.368 }, 00:08:34.368 { 00:08:34.368 "name": "pt2", 00:08:34.368 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:34.368 "is_configured": true, 00:08:34.368 "data_offset": 2048, 00:08:34.368 "data_size": 63488 00:08:34.368 }, 00:08:34.368 { 00:08:34.368 "name": "pt3", 00:08:34.368 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:34.368 "is_configured": true, 00:08:34.368 "data_offset": 2048, 00:08:34.368 "data_size": 63488 00:08:34.368 } 00:08:34.368 ] 00:08:34.368 } 00:08:34.368 } 00:08:34.368 }' 00:08:34.368 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:34.368 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:34.368 pt2 00:08:34.368 pt3' 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.369 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.369 [2024-12-07 05:36:07.716160] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:34.629 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.629 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=fb97627a-712b-4bd2-833c-7ed0811a12e1 00:08:34.629 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z fb97627a-712b-4bd2-833c-7ed0811a12e1 ']' 00:08:34.629 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:34.629 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.629 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.629 [2024-12-07 05:36:07.759852] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:34.629 [2024-12-07 05:36:07.759886] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:34.630 [2024-12-07 05:36:07.759972] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:34.630 [2024-12-07 05:36:07.760043] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:34.630 [2024-12-07 05:36:07.760062] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.630 [2024-12-07 05:36:07.875700] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:34.630 [2024-12-07 05:36:07.877573] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:34.630 [2024-12-07 05:36:07.877619] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:34.630 [2024-12-07 05:36:07.877678] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:34.630 [2024-12-07 05:36:07.877741] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:34.630 [2024-12-07 05:36:07.877771] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:34.630 [2024-12-07 05:36:07.877784] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:34.630 [2024-12-07 05:36:07.877795] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:34.630 request: 00:08:34.630 { 00:08:34.630 "name": "raid_bdev1", 00:08:34.630 "raid_level": "concat", 00:08:34.630 "base_bdevs": [ 00:08:34.630 "malloc1", 00:08:34.630 "malloc2", 00:08:34.630 "malloc3" 00:08:34.630 ], 00:08:34.630 "strip_size_kb": 64, 00:08:34.630 "superblock": false, 00:08:34.630 "method": "bdev_raid_create", 00:08:34.630 "req_id": 1 00:08:34.630 } 00:08:34.630 Got JSON-RPC error response 00:08:34.630 response: 00:08:34.630 { 00:08:34.630 "code": -17, 00:08:34.630 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:34.630 } 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.630 [2024-12-07 05:36:07.935550] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:34.630 [2024-12-07 05:36:07.935654] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:34.630 [2024-12-07 05:36:07.935675] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:34.630 [2024-12-07 05:36:07.935686] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:34.630 [2024-12-07 05:36:07.937855] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:34.630 [2024-12-07 05:36:07.937892] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:34.630 [2024-12-07 05:36:07.937960] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:34.630 [2024-12-07 05:36:07.938008] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:34.630 pt1 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:34.630 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.631 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.631 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.631 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:34.631 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.631 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:34.631 "name": "raid_bdev1", 00:08:34.631 "uuid": "fb97627a-712b-4bd2-833c-7ed0811a12e1", 00:08:34.631 "strip_size_kb": 64, 00:08:34.631 "state": "configuring", 00:08:34.631 "raid_level": "concat", 00:08:34.631 "superblock": true, 00:08:34.631 "num_base_bdevs": 3, 00:08:34.631 "num_base_bdevs_discovered": 1, 00:08:34.631 "num_base_bdevs_operational": 3, 00:08:34.631 "base_bdevs_list": [ 00:08:34.631 { 00:08:34.631 "name": "pt1", 00:08:34.631 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:34.631 "is_configured": true, 00:08:34.631 "data_offset": 2048, 00:08:34.631 "data_size": 63488 00:08:34.631 }, 00:08:34.631 { 00:08:34.631 "name": null, 00:08:34.631 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:34.631 "is_configured": false, 00:08:34.631 "data_offset": 2048, 00:08:34.631 "data_size": 63488 00:08:34.631 }, 00:08:34.631 { 00:08:34.631 "name": null, 00:08:34.631 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:34.631 "is_configured": false, 00:08:34.631 "data_offset": 2048, 00:08:34.631 "data_size": 63488 00:08:34.631 } 00:08:34.631 ] 00:08:34.631 }' 00:08:34.631 05:36:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:34.631 05:36:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.200 [2024-12-07 05:36:08.374831] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:35.200 [2024-12-07 05:36:08.374944] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:35.200 [2024-12-07 05:36:08.374970] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:35.200 [2024-12-07 05:36:08.374985] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:35.200 [2024-12-07 05:36:08.375389] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:35.200 [2024-12-07 05:36:08.375418] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:35.200 [2024-12-07 05:36:08.375491] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:35.200 [2024-12-07 05:36:08.375514] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:35.200 pt2 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.200 [2024-12-07 05:36:08.382817] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.200 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:35.200 "name": "raid_bdev1", 00:08:35.200 "uuid": "fb97627a-712b-4bd2-833c-7ed0811a12e1", 00:08:35.200 "strip_size_kb": 64, 00:08:35.200 "state": "configuring", 00:08:35.200 "raid_level": "concat", 00:08:35.200 "superblock": true, 00:08:35.200 "num_base_bdevs": 3, 00:08:35.200 "num_base_bdevs_discovered": 1, 00:08:35.200 "num_base_bdevs_operational": 3, 00:08:35.200 "base_bdevs_list": [ 00:08:35.200 { 00:08:35.201 "name": "pt1", 00:08:35.201 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:35.201 "is_configured": true, 00:08:35.201 "data_offset": 2048, 00:08:35.201 "data_size": 63488 00:08:35.201 }, 00:08:35.201 { 00:08:35.201 "name": null, 00:08:35.201 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:35.201 "is_configured": false, 00:08:35.201 "data_offset": 0, 00:08:35.201 "data_size": 63488 00:08:35.201 }, 00:08:35.201 { 00:08:35.201 "name": null, 00:08:35.201 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:35.201 "is_configured": false, 00:08:35.201 "data_offset": 2048, 00:08:35.201 "data_size": 63488 00:08:35.201 } 00:08:35.201 ] 00:08:35.201 }' 00:08:35.201 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:35.201 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.461 [2024-12-07 05:36:08.798111] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:35.461 [2024-12-07 05:36:08.798173] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:35.461 [2024-12-07 05:36:08.798195] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:35.461 [2024-12-07 05:36:08.798205] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:35.461 [2024-12-07 05:36:08.798655] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:35.461 [2024-12-07 05:36:08.798680] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:35.461 [2024-12-07 05:36:08.798759] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:35.461 [2024-12-07 05:36:08.798819] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:35.461 pt2 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.461 [2024-12-07 05:36:08.806092] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:35.461 [2024-12-07 05:36:08.806184] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:35.461 [2024-12-07 05:36:08.806208] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:35.461 [2024-12-07 05:36:08.806218] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:35.461 [2024-12-07 05:36:08.806587] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:35.461 [2024-12-07 05:36:08.806619] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:35.461 [2024-12-07 05:36:08.806694] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:35.461 [2024-12-07 05:36:08.806714] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:35.461 [2024-12-07 05:36:08.806813] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:35.461 [2024-12-07 05:36:08.806822] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:35.461 [2024-12-07 05:36:08.807069] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:35.461 [2024-12-07 05:36:08.807200] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:35.461 [2024-12-07 05:36:08.807214] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:35.461 [2024-12-07 05:36:08.807319] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:35.461 pt3 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.461 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:35.721 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.721 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:35.721 "name": "raid_bdev1", 00:08:35.721 "uuid": "fb97627a-712b-4bd2-833c-7ed0811a12e1", 00:08:35.721 "strip_size_kb": 64, 00:08:35.721 "state": "online", 00:08:35.721 "raid_level": "concat", 00:08:35.721 "superblock": true, 00:08:35.721 "num_base_bdevs": 3, 00:08:35.721 "num_base_bdevs_discovered": 3, 00:08:35.721 "num_base_bdevs_operational": 3, 00:08:35.721 "base_bdevs_list": [ 00:08:35.721 { 00:08:35.721 "name": "pt1", 00:08:35.721 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:35.721 "is_configured": true, 00:08:35.721 "data_offset": 2048, 00:08:35.721 "data_size": 63488 00:08:35.721 }, 00:08:35.721 { 00:08:35.721 "name": "pt2", 00:08:35.721 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:35.721 "is_configured": true, 00:08:35.721 "data_offset": 2048, 00:08:35.721 "data_size": 63488 00:08:35.721 }, 00:08:35.721 { 00:08:35.721 "name": "pt3", 00:08:35.721 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:35.721 "is_configured": true, 00:08:35.721 "data_offset": 2048, 00:08:35.721 "data_size": 63488 00:08:35.721 } 00:08:35.721 ] 00:08:35.721 }' 00:08:35.721 05:36:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:35.721 05:36:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.980 [2024-12-07 05:36:09.261622] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.980 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:35.980 "name": "raid_bdev1", 00:08:35.980 "aliases": [ 00:08:35.980 "fb97627a-712b-4bd2-833c-7ed0811a12e1" 00:08:35.980 ], 00:08:35.980 "product_name": "Raid Volume", 00:08:35.980 "block_size": 512, 00:08:35.980 "num_blocks": 190464, 00:08:35.980 "uuid": "fb97627a-712b-4bd2-833c-7ed0811a12e1", 00:08:35.980 "assigned_rate_limits": { 00:08:35.980 "rw_ios_per_sec": 0, 00:08:35.980 "rw_mbytes_per_sec": 0, 00:08:35.981 "r_mbytes_per_sec": 0, 00:08:35.981 "w_mbytes_per_sec": 0 00:08:35.981 }, 00:08:35.981 "claimed": false, 00:08:35.981 "zoned": false, 00:08:35.981 "supported_io_types": { 00:08:35.981 "read": true, 00:08:35.981 "write": true, 00:08:35.981 "unmap": true, 00:08:35.981 "flush": true, 00:08:35.981 "reset": true, 00:08:35.981 "nvme_admin": false, 00:08:35.981 "nvme_io": false, 00:08:35.981 "nvme_io_md": false, 00:08:35.981 "write_zeroes": true, 00:08:35.981 "zcopy": false, 00:08:35.981 "get_zone_info": false, 00:08:35.981 "zone_management": false, 00:08:35.981 "zone_append": false, 00:08:35.981 "compare": false, 00:08:35.981 "compare_and_write": false, 00:08:35.981 "abort": false, 00:08:35.981 "seek_hole": false, 00:08:35.981 "seek_data": false, 00:08:35.981 "copy": false, 00:08:35.981 "nvme_iov_md": false 00:08:35.981 }, 00:08:35.981 "memory_domains": [ 00:08:35.981 { 00:08:35.981 "dma_device_id": "system", 00:08:35.981 "dma_device_type": 1 00:08:35.981 }, 00:08:35.981 { 00:08:35.981 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:35.981 "dma_device_type": 2 00:08:35.981 }, 00:08:35.981 { 00:08:35.981 "dma_device_id": "system", 00:08:35.981 "dma_device_type": 1 00:08:35.981 }, 00:08:35.981 { 00:08:35.981 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:35.981 "dma_device_type": 2 00:08:35.981 }, 00:08:35.981 { 00:08:35.981 "dma_device_id": "system", 00:08:35.981 "dma_device_type": 1 00:08:35.981 }, 00:08:35.981 { 00:08:35.981 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:35.981 "dma_device_type": 2 00:08:35.981 } 00:08:35.981 ], 00:08:35.981 "driver_specific": { 00:08:35.981 "raid": { 00:08:35.981 "uuid": "fb97627a-712b-4bd2-833c-7ed0811a12e1", 00:08:35.981 "strip_size_kb": 64, 00:08:35.981 "state": "online", 00:08:35.981 "raid_level": "concat", 00:08:35.981 "superblock": true, 00:08:35.981 "num_base_bdevs": 3, 00:08:35.981 "num_base_bdevs_discovered": 3, 00:08:35.981 "num_base_bdevs_operational": 3, 00:08:35.981 "base_bdevs_list": [ 00:08:35.981 { 00:08:35.981 "name": "pt1", 00:08:35.981 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:35.981 "is_configured": true, 00:08:35.981 "data_offset": 2048, 00:08:35.981 "data_size": 63488 00:08:35.981 }, 00:08:35.981 { 00:08:35.981 "name": "pt2", 00:08:35.981 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:35.981 "is_configured": true, 00:08:35.981 "data_offset": 2048, 00:08:35.981 "data_size": 63488 00:08:35.981 }, 00:08:35.981 { 00:08:35.981 "name": "pt3", 00:08:35.981 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:35.981 "is_configured": true, 00:08:35.981 "data_offset": 2048, 00:08:35.981 "data_size": 63488 00:08:35.981 } 00:08:35.981 ] 00:08:35.981 } 00:08:35.981 } 00:08:35.981 }' 00:08:35.981 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:35.981 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:35.981 pt2 00:08:35.981 pt3' 00:08:35.981 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:36.240 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.241 [2024-12-07 05:36:09.525117] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' fb97627a-712b-4bd2-833c-7ed0811a12e1 '!=' fb97627a-712b-4bd2-833c-7ed0811a12e1 ']' 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 77616 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 77616 ']' 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 77616 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77616 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77616' 00:08:36.241 killing process with pid 77616 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 77616 00:08:36.241 [2024-12-07 05:36:09.603499] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:36.241 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 77616 00:08:36.241 [2024-12-07 05:36:09.603639] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:36.241 [2024-12-07 05:36:09.603737] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:36.241 [2024-12-07 05:36:09.603788] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:36.500 [2024-12-07 05:36:09.637898] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:36.500 05:36:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:36.500 00:08:36.500 real 0m3.892s 00:08:36.500 user 0m6.227s 00:08:36.500 sys 0m0.762s 00:08:36.500 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:36.500 05:36:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.500 ************************************ 00:08:36.500 END TEST raid_superblock_test 00:08:36.500 ************************************ 00:08:36.758 05:36:09 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:08:36.758 05:36:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:36.758 05:36:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:36.758 05:36:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:36.758 ************************************ 00:08:36.758 START TEST raid_read_error_test 00:08:36.758 ************************************ 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 read 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.u7u2Aux8Oi 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=77853 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 77853 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 77853 ']' 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:36.758 05:36:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:36.759 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:36.759 05:36:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:36.759 05:36:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:36.759 05:36:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.759 [2024-12-07 05:36:10.005134] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:08:36.759 [2024-12-07 05:36:10.005257] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77853 ] 00:08:37.050 [2024-12-07 05:36:10.159766] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:37.050 [2024-12-07 05:36:10.185533] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:37.050 [2024-12-07 05:36:10.228816] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:37.050 [2024-12-07 05:36:10.228851] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.616 BaseBdev1_malloc 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.616 true 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.616 [2024-12-07 05:36:10.861069] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:37.616 [2024-12-07 05:36:10.861134] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:37.616 [2024-12-07 05:36:10.861155] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:37.616 [2024-12-07 05:36:10.861163] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:37.616 [2024-12-07 05:36:10.863556] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:37.616 [2024-12-07 05:36:10.863662] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:37.616 BaseBdev1 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.616 BaseBdev2_malloc 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.616 true 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.616 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.616 [2024-12-07 05:36:10.889891] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:37.616 [2024-12-07 05:36:10.889987] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:37.617 [2024-12-07 05:36:10.890009] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:37.617 [2024-12-07 05:36:10.890027] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:37.617 [2024-12-07 05:36:10.892273] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:37.617 [2024-12-07 05:36:10.892309] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:37.617 BaseBdev2 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.617 BaseBdev3_malloc 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.617 true 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.617 [2024-12-07 05:36:10.918567] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:37.617 [2024-12-07 05:36:10.918615] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:37.617 [2024-12-07 05:36:10.918648] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:37.617 [2024-12-07 05:36:10.918657] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:37.617 [2024-12-07 05:36:10.920877] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:37.617 [2024-12-07 05:36:10.920912] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:37.617 BaseBdev3 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.617 [2024-12-07 05:36:10.926626] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:37.617 [2024-12-07 05:36:10.928489] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:37.617 [2024-12-07 05:36:10.928566] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:37.617 [2024-12-07 05:36:10.928760] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:37.617 [2024-12-07 05:36:10.928775] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:37.617 [2024-12-07 05:36:10.929017] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:08:37.617 [2024-12-07 05:36:10.929154] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:37.617 [2024-12-07 05:36:10.929171] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:37.617 [2024-12-07 05:36:10.929296] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:37.617 "name": "raid_bdev1", 00:08:37.617 "uuid": "f0f8a4bc-92ad-460b-b9d1-cfe4ae452e6a", 00:08:37.617 "strip_size_kb": 64, 00:08:37.617 "state": "online", 00:08:37.617 "raid_level": "concat", 00:08:37.617 "superblock": true, 00:08:37.617 "num_base_bdevs": 3, 00:08:37.617 "num_base_bdevs_discovered": 3, 00:08:37.617 "num_base_bdevs_operational": 3, 00:08:37.617 "base_bdevs_list": [ 00:08:37.617 { 00:08:37.617 "name": "BaseBdev1", 00:08:37.617 "uuid": "5736b1df-7981-5494-831b-fb039532242c", 00:08:37.617 "is_configured": true, 00:08:37.617 "data_offset": 2048, 00:08:37.617 "data_size": 63488 00:08:37.617 }, 00:08:37.617 { 00:08:37.617 "name": "BaseBdev2", 00:08:37.617 "uuid": "b6f21c91-3a5c-5308-bcac-159f05ee7f4c", 00:08:37.617 "is_configured": true, 00:08:37.617 "data_offset": 2048, 00:08:37.617 "data_size": 63488 00:08:37.617 }, 00:08:37.617 { 00:08:37.617 "name": "BaseBdev3", 00:08:37.617 "uuid": "08e83e28-9566-5ff7-b16a-6ce7bb7400fd", 00:08:37.617 "is_configured": true, 00:08:37.617 "data_offset": 2048, 00:08:37.617 "data_size": 63488 00:08:37.617 } 00:08:37.617 ] 00:08:37.617 }' 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:37.617 05:36:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.184 05:36:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:38.184 05:36:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:38.184 [2024-12-07 05:36:11.438167] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:39.118 "name": "raid_bdev1", 00:08:39.118 "uuid": "f0f8a4bc-92ad-460b-b9d1-cfe4ae452e6a", 00:08:39.118 "strip_size_kb": 64, 00:08:39.118 "state": "online", 00:08:39.118 "raid_level": "concat", 00:08:39.118 "superblock": true, 00:08:39.118 "num_base_bdevs": 3, 00:08:39.118 "num_base_bdevs_discovered": 3, 00:08:39.118 "num_base_bdevs_operational": 3, 00:08:39.118 "base_bdevs_list": [ 00:08:39.118 { 00:08:39.118 "name": "BaseBdev1", 00:08:39.118 "uuid": "5736b1df-7981-5494-831b-fb039532242c", 00:08:39.118 "is_configured": true, 00:08:39.118 "data_offset": 2048, 00:08:39.118 "data_size": 63488 00:08:39.118 }, 00:08:39.118 { 00:08:39.118 "name": "BaseBdev2", 00:08:39.118 "uuid": "b6f21c91-3a5c-5308-bcac-159f05ee7f4c", 00:08:39.118 "is_configured": true, 00:08:39.118 "data_offset": 2048, 00:08:39.118 "data_size": 63488 00:08:39.118 }, 00:08:39.118 { 00:08:39.118 "name": "BaseBdev3", 00:08:39.118 "uuid": "08e83e28-9566-5ff7-b16a-6ce7bb7400fd", 00:08:39.118 "is_configured": true, 00:08:39.118 "data_offset": 2048, 00:08:39.118 "data_size": 63488 00:08:39.118 } 00:08:39.118 ] 00:08:39.118 }' 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:39.118 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.683 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:39.683 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.683 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.683 [2024-12-07 05:36:12.810478] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:39.683 [2024-12-07 05:36:12.810580] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:39.683 [2024-12-07 05:36:12.813343] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:39.683 [2024-12-07 05:36:12.813433] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:39.683 [2024-12-07 05:36:12.813502] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:39.683 [2024-12-07 05:36:12.813561] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:39.683 { 00:08:39.683 "results": [ 00:08:39.683 { 00:08:39.683 "job": "raid_bdev1", 00:08:39.683 "core_mask": "0x1", 00:08:39.683 "workload": "randrw", 00:08:39.683 "percentage": 50, 00:08:39.683 "status": "finished", 00:08:39.683 "queue_depth": 1, 00:08:39.683 "io_size": 131072, 00:08:39.683 "runtime": 1.373311, 00:08:39.683 "iops": 15685.449253665047, 00:08:39.684 "mibps": 1960.6811567081309, 00:08:39.684 "io_failed": 1, 00:08:39.684 "io_timeout": 0, 00:08:39.684 "avg_latency_us": 87.94339839428127, 00:08:39.684 "min_latency_us": 27.276855895196505, 00:08:39.684 "max_latency_us": 1552.5449781659388 00:08:39.684 } 00:08:39.684 ], 00:08:39.684 "core_count": 1 00:08:39.684 } 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 77853 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 77853 ']' 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 77853 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77853 00:08:39.684 killing process with pid 77853 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77853' 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 77853 00:08:39.684 05:36:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 77853 00:08:39.684 [2024-12-07 05:36:12.845028] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:39.684 [2024-12-07 05:36:12.871435] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:39.942 05:36:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.u7u2Aux8Oi 00:08:39.942 05:36:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:39.942 05:36:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:39.942 ************************************ 00:08:39.942 END TEST raid_read_error_test 00:08:39.942 ************************************ 00:08:39.942 05:36:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:08:39.942 05:36:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:39.942 05:36:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:39.942 05:36:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:39.942 05:36:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:08:39.942 00:08:39.942 real 0m3.174s 00:08:39.942 user 0m4.031s 00:08:39.942 sys 0m0.501s 00:08:39.942 05:36:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:39.942 05:36:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.942 05:36:13 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:08:39.942 05:36:13 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:39.942 05:36:13 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:39.942 05:36:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:39.942 ************************************ 00:08:39.942 START TEST raid_write_error_test 00:08:39.942 ************************************ 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 write 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.78iikG7TZo 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=77982 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 77982 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 77982 ']' 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:39.942 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.942 05:36:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:39.942 [2024-12-07 05:36:13.239843] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:08:39.942 [2024-12-07 05:36:13.240049] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77982 ] 00:08:40.201 [2024-12-07 05:36:13.394809] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:40.201 [2024-12-07 05:36:13.420294] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:40.201 [2024-12-07 05:36:13.462560] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:40.201 [2024-12-07 05:36:13.462596] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.769 BaseBdev1_malloc 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.769 true 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.769 [2024-12-07 05:36:14.101920] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:40.769 [2024-12-07 05:36:14.101995] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:40.769 [2024-12-07 05:36:14.102018] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:40.769 [2024-12-07 05:36:14.102027] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:40.769 [2024-12-07 05:36:14.104203] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:40.769 [2024-12-07 05:36:14.104303] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:40.769 BaseBdev1 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.769 BaseBdev2_malloc 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.769 true 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.769 [2024-12-07 05:36:14.130460] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:40.769 [2024-12-07 05:36:14.130511] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:40.769 [2024-12-07 05:36:14.130545] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:40.769 [2024-12-07 05:36:14.130564] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:40.769 [2024-12-07 05:36:14.132927] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:40.769 [2024-12-07 05:36:14.132975] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:40.769 BaseBdev2 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.769 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.030 BaseBdev3_malloc 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.030 true 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.030 [2024-12-07 05:36:14.159370] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:41.030 [2024-12-07 05:36:14.159421] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:41.030 [2024-12-07 05:36:14.159440] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:41.030 [2024-12-07 05:36:14.159449] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:41.030 [2024-12-07 05:36:14.161592] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:41.030 [2024-12-07 05:36:14.161642] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:41.030 BaseBdev3 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.030 [2024-12-07 05:36:14.167418] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:41.030 [2024-12-07 05:36:14.169306] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:41.030 [2024-12-07 05:36:14.169376] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:41.030 [2024-12-07 05:36:14.169546] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:41.030 [2024-12-07 05:36:14.169560] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:41.030 [2024-12-07 05:36:14.169843] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:08:41.030 [2024-12-07 05:36:14.169979] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:41.030 [2024-12-07 05:36:14.169995] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:41.030 [2024-12-07 05:36:14.170126] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.030 "name": "raid_bdev1", 00:08:41.030 "uuid": "7388ce4b-b2ca-4e38-a013-07ecee9c6130", 00:08:41.030 "strip_size_kb": 64, 00:08:41.030 "state": "online", 00:08:41.030 "raid_level": "concat", 00:08:41.030 "superblock": true, 00:08:41.030 "num_base_bdevs": 3, 00:08:41.030 "num_base_bdevs_discovered": 3, 00:08:41.030 "num_base_bdevs_operational": 3, 00:08:41.030 "base_bdevs_list": [ 00:08:41.030 { 00:08:41.030 "name": "BaseBdev1", 00:08:41.030 "uuid": "e29c37d9-b67b-5e5f-b223-cc5c88f28bc2", 00:08:41.030 "is_configured": true, 00:08:41.030 "data_offset": 2048, 00:08:41.030 "data_size": 63488 00:08:41.030 }, 00:08:41.030 { 00:08:41.030 "name": "BaseBdev2", 00:08:41.030 "uuid": "a891e1ca-7462-5bc1-a837-7000a8017cdd", 00:08:41.030 "is_configured": true, 00:08:41.030 "data_offset": 2048, 00:08:41.030 "data_size": 63488 00:08:41.030 }, 00:08:41.030 { 00:08:41.030 "name": "BaseBdev3", 00:08:41.030 "uuid": "91f0a1bc-58a2-56c0-85b7-2c14dd83be74", 00:08:41.030 "is_configured": true, 00:08:41.030 "data_offset": 2048, 00:08:41.030 "data_size": 63488 00:08:41.030 } 00:08:41.030 ] 00:08:41.030 }' 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.030 05:36:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.289 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:41.289 05:36:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:41.549 [2024-12-07 05:36:14.679010] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.486 "name": "raid_bdev1", 00:08:42.486 "uuid": "7388ce4b-b2ca-4e38-a013-07ecee9c6130", 00:08:42.486 "strip_size_kb": 64, 00:08:42.486 "state": "online", 00:08:42.486 "raid_level": "concat", 00:08:42.486 "superblock": true, 00:08:42.486 "num_base_bdevs": 3, 00:08:42.486 "num_base_bdevs_discovered": 3, 00:08:42.486 "num_base_bdevs_operational": 3, 00:08:42.486 "base_bdevs_list": [ 00:08:42.486 { 00:08:42.486 "name": "BaseBdev1", 00:08:42.486 "uuid": "e29c37d9-b67b-5e5f-b223-cc5c88f28bc2", 00:08:42.486 "is_configured": true, 00:08:42.486 "data_offset": 2048, 00:08:42.486 "data_size": 63488 00:08:42.486 }, 00:08:42.486 { 00:08:42.486 "name": "BaseBdev2", 00:08:42.486 "uuid": "a891e1ca-7462-5bc1-a837-7000a8017cdd", 00:08:42.486 "is_configured": true, 00:08:42.486 "data_offset": 2048, 00:08:42.486 "data_size": 63488 00:08:42.486 }, 00:08:42.486 { 00:08:42.486 "name": "BaseBdev3", 00:08:42.486 "uuid": "91f0a1bc-58a2-56c0-85b7-2c14dd83be74", 00:08:42.486 "is_configured": true, 00:08:42.486 "data_offset": 2048, 00:08:42.486 "data_size": 63488 00:08:42.486 } 00:08:42.486 ] 00:08:42.486 }' 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.486 05:36:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.745 05:36:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:42.745 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.745 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.745 [2024-12-07 05:36:16.043084] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:42.745 [2024-12-07 05:36:16.043124] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:42.745 [2024-12-07 05:36:16.045811] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:42.745 [2024-12-07 05:36:16.045866] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:42.745 [2024-12-07 05:36:16.045902] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:42.745 [2024-12-07 05:36:16.045915] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:42.745 { 00:08:42.745 "results": [ 00:08:42.745 { 00:08:42.745 "job": "raid_bdev1", 00:08:42.746 "core_mask": "0x1", 00:08:42.746 "workload": "randrw", 00:08:42.746 "percentage": 50, 00:08:42.746 "status": "finished", 00:08:42.746 "queue_depth": 1, 00:08:42.746 "io_size": 131072, 00:08:42.746 "runtime": 1.364887, 00:08:42.746 "iops": 15715.586711573926, 00:08:42.746 "mibps": 1964.4483389467407, 00:08:42.746 "io_failed": 1, 00:08:42.746 "io_timeout": 0, 00:08:42.746 "avg_latency_us": 87.82259997854356, 00:08:42.746 "min_latency_us": 27.165065502183406, 00:08:42.746 "max_latency_us": 1466.6899563318777 00:08:42.746 } 00:08:42.746 ], 00:08:42.746 "core_count": 1 00:08:42.746 } 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 77982 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 77982 ']' 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 77982 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77982 00:08:42.746 killing process with pid 77982 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77982' 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 77982 00:08:42.746 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 77982 00:08:42.746 [2024-12-07 05:36:16.089233] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:43.006 [2024-12-07 05:36:16.115390] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:43.006 05:36:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.78iikG7TZo 00:08:43.006 05:36:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:43.006 05:36:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:43.006 05:36:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:08:43.006 05:36:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:43.006 05:36:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:43.006 05:36:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:43.006 05:36:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:08:43.006 00:08:43.006 real 0m3.187s 00:08:43.006 user 0m4.067s 00:08:43.006 sys 0m0.496s 00:08:43.006 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:43.006 05:36:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.006 ************************************ 00:08:43.006 END TEST raid_write_error_test 00:08:43.006 ************************************ 00:08:43.266 05:36:16 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:43.266 05:36:16 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:08:43.266 05:36:16 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:43.266 05:36:16 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:43.266 05:36:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:43.266 ************************************ 00:08:43.266 START TEST raid_state_function_test 00:08:43.266 ************************************ 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 false 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:43.266 Process raid pid: 78115 00:08:43.266 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=78115 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78115' 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 78115 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 78115 ']' 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.266 05:36:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:43.266 [2024-12-07 05:36:16.494956] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:08:43.266 [2024-12-07 05:36:16.495155] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:43.266 [2024-12-07 05:36:16.630506] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:43.525 [2024-12-07 05:36:16.656837] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:43.525 [2024-12-07 05:36:16.699615] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:43.525 [2024-12-07 05:36:16.699670] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.095 [2024-12-07 05:36:17.322169] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:44.095 [2024-12-07 05:36:17.322231] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:44.095 [2024-12-07 05:36:17.322257] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:44.095 [2024-12-07 05:36:17.322267] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:44.095 [2024-12-07 05:36:17.322273] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:44.095 [2024-12-07 05:36:17.322286] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.095 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.095 "name": "Existed_Raid", 00:08:44.095 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.095 "strip_size_kb": 0, 00:08:44.095 "state": "configuring", 00:08:44.095 "raid_level": "raid1", 00:08:44.095 "superblock": false, 00:08:44.095 "num_base_bdevs": 3, 00:08:44.095 "num_base_bdevs_discovered": 0, 00:08:44.095 "num_base_bdevs_operational": 3, 00:08:44.095 "base_bdevs_list": [ 00:08:44.095 { 00:08:44.095 "name": "BaseBdev1", 00:08:44.095 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.095 "is_configured": false, 00:08:44.095 "data_offset": 0, 00:08:44.095 "data_size": 0 00:08:44.095 }, 00:08:44.095 { 00:08:44.095 "name": "BaseBdev2", 00:08:44.095 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.095 "is_configured": false, 00:08:44.095 "data_offset": 0, 00:08:44.095 "data_size": 0 00:08:44.095 }, 00:08:44.095 { 00:08:44.095 "name": "BaseBdev3", 00:08:44.095 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.096 "is_configured": false, 00:08:44.096 "data_offset": 0, 00:08:44.096 "data_size": 0 00:08:44.096 } 00:08:44.096 ] 00:08:44.096 }' 00:08:44.096 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.096 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.355 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:44.355 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.355 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.355 [2024-12-07 05:36:17.705450] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:44.355 [2024-12-07 05:36:17.705500] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:44.355 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.355 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:44.355 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.355 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.355 [2024-12-07 05:36:17.713442] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:44.355 [2024-12-07 05:36:17.713487] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:44.355 [2024-12-07 05:36:17.713496] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:44.355 [2024-12-07 05:36:17.713521] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:44.355 [2024-12-07 05:36:17.713527] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:44.355 [2024-12-07 05:36:17.713536] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:44.355 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.355 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:44.355 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.355 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.614 BaseBdev1 00:08:44.614 [2024-12-07 05:36:17.730213] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.614 [ 00:08:44.614 { 00:08:44.614 "name": "BaseBdev1", 00:08:44.614 "aliases": [ 00:08:44.614 "64bd938c-32e3-485d-a66f-4fd459c0db01" 00:08:44.614 ], 00:08:44.614 "product_name": "Malloc disk", 00:08:44.614 "block_size": 512, 00:08:44.614 "num_blocks": 65536, 00:08:44.614 "uuid": "64bd938c-32e3-485d-a66f-4fd459c0db01", 00:08:44.614 "assigned_rate_limits": { 00:08:44.614 "rw_ios_per_sec": 0, 00:08:44.614 "rw_mbytes_per_sec": 0, 00:08:44.614 "r_mbytes_per_sec": 0, 00:08:44.614 "w_mbytes_per_sec": 0 00:08:44.614 }, 00:08:44.614 "claimed": true, 00:08:44.614 "claim_type": "exclusive_write", 00:08:44.614 "zoned": false, 00:08:44.614 "supported_io_types": { 00:08:44.614 "read": true, 00:08:44.614 "write": true, 00:08:44.614 "unmap": true, 00:08:44.614 "flush": true, 00:08:44.614 "reset": true, 00:08:44.614 "nvme_admin": false, 00:08:44.614 "nvme_io": false, 00:08:44.614 "nvme_io_md": false, 00:08:44.614 "write_zeroes": true, 00:08:44.614 "zcopy": true, 00:08:44.614 "get_zone_info": false, 00:08:44.614 "zone_management": false, 00:08:44.614 "zone_append": false, 00:08:44.614 "compare": false, 00:08:44.614 "compare_and_write": false, 00:08:44.614 "abort": true, 00:08:44.614 "seek_hole": false, 00:08:44.614 "seek_data": false, 00:08:44.614 "copy": true, 00:08:44.614 "nvme_iov_md": false 00:08:44.614 }, 00:08:44.614 "memory_domains": [ 00:08:44.614 { 00:08:44.614 "dma_device_id": "system", 00:08:44.614 "dma_device_type": 1 00:08:44.614 }, 00:08:44.614 { 00:08:44.614 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:44.614 "dma_device_type": 2 00:08:44.614 } 00:08:44.614 ], 00:08:44.614 "driver_specific": {} 00:08:44.614 } 00:08:44.614 ] 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.614 "name": "Existed_Raid", 00:08:44.614 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.614 "strip_size_kb": 0, 00:08:44.614 "state": "configuring", 00:08:44.614 "raid_level": "raid1", 00:08:44.614 "superblock": false, 00:08:44.614 "num_base_bdevs": 3, 00:08:44.614 "num_base_bdevs_discovered": 1, 00:08:44.614 "num_base_bdevs_operational": 3, 00:08:44.614 "base_bdevs_list": [ 00:08:44.614 { 00:08:44.614 "name": "BaseBdev1", 00:08:44.614 "uuid": "64bd938c-32e3-485d-a66f-4fd459c0db01", 00:08:44.614 "is_configured": true, 00:08:44.614 "data_offset": 0, 00:08:44.614 "data_size": 65536 00:08:44.614 }, 00:08:44.614 { 00:08:44.614 "name": "BaseBdev2", 00:08:44.614 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.614 "is_configured": false, 00:08:44.614 "data_offset": 0, 00:08:44.614 "data_size": 0 00:08:44.614 }, 00:08:44.614 { 00:08:44.614 "name": "BaseBdev3", 00:08:44.614 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.614 "is_configured": false, 00:08:44.614 "data_offset": 0, 00:08:44.614 "data_size": 0 00:08:44.614 } 00:08:44.614 ] 00:08:44.614 }' 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.614 05:36:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.872 [2024-12-07 05:36:18.129578] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:44.872 [2024-12-07 05:36:18.129698] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.872 [2024-12-07 05:36:18.137604] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:44.872 [2024-12-07 05:36:18.139591] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:44.872 [2024-12-07 05:36:18.139683] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:44.872 [2024-12-07 05:36:18.139718] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:44.872 [2024-12-07 05:36:18.139775] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.872 "name": "Existed_Raid", 00:08:44.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.872 "strip_size_kb": 0, 00:08:44.872 "state": "configuring", 00:08:44.872 "raid_level": "raid1", 00:08:44.872 "superblock": false, 00:08:44.872 "num_base_bdevs": 3, 00:08:44.872 "num_base_bdevs_discovered": 1, 00:08:44.872 "num_base_bdevs_operational": 3, 00:08:44.872 "base_bdevs_list": [ 00:08:44.872 { 00:08:44.872 "name": "BaseBdev1", 00:08:44.872 "uuid": "64bd938c-32e3-485d-a66f-4fd459c0db01", 00:08:44.872 "is_configured": true, 00:08:44.872 "data_offset": 0, 00:08:44.872 "data_size": 65536 00:08:44.872 }, 00:08:44.872 { 00:08:44.872 "name": "BaseBdev2", 00:08:44.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.872 "is_configured": false, 00:08:44.872 "data_offset": 0, 00:08:44.872 "data_size": 0 00:08:44.872 }, 00:08:44.872 { 00:08:44.872 "name": "BaseBdev3", 00:08:44.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.872 "is_configured": false, 00:08:44.872 "data_offset": 0, 00:08:44.872 "data_size": 0 00:08:44.872 } 00:08:44.872 ] 00:08:44.872 }' 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.872 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.439 [2024-12-07 05:36:18.599826] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:45.439 BaseBdev2 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.439 [ 00:08:45.439 { 00:08:45.439 "name": "BaseBdev2", 00:08:45.439 "aliases": [ 00:08:45.439 "ea9f4306-89c3-4043-9763-a2cfa076d3c8" 00:08:45.439 ], 00:08:45.439 "product_name": "Malloc disk", 00:08:45.439 "block_size": 512, 00:08:45.439 "num_blocks": 65536, 00:08:45.439 "uuid": "ea9f4306-89c3-4043-9763-a2cfa076d3c8", 00:08:45.439 "assigned_rate_limits": { 00:08:45.439 "rw_ios_per_sec": 0, 00:08:45.439 "rw_mbytes_per_sec": 0, 00:08:45.439 "r_mbytes_per_sec": 0, 00:08:45.439 "w_mbytes_per_sec": 0 00:08:45.439 }, 00:08:45.439 "claimed": true, 00:08:45.439 "claim_type": "exclusive_write", 00:08:45.439 "zoned": false, 00:08:45.439 "supported_io_types": { 00:08:45.439 "read": true, 00:08:45.439 "write": true, 00:08:45.439 "unmap": true, 00:08:45.439 "flush": true, 00:08:45.439 "reset": true, 00:08:45.439 "nvme_admin": false, 00:08:45.439 "nvme_io": false, 00:08:45.439 "nvme_io_md": false, 00:08:45.439 "write_zeroes": true, 00:08:45.439 "zcopy": true, 00:08:45.439 "get_zone_info": false, 00:08:45.439 "zone_management": false, 00:08:45.439 "zone_append": false, 00:08:45.439 "compare": false, 00:08:45.439 "compare_and_write": false, 00:08:45.439 "abort": true, 00:08:45.439 "seek_hole": false, 00:08:45.439 "seek_data": false, 00:08:45.439 "copy": true, 00:08:45.439 "nvme_iov_md": false 00:08:45.439 }, 00:08:45.439 "memory_domains": [ 00:08:45.439 { 00:08:45.439 "dma_device_id": "system", 00:08:45.439 "dma_device_type": 1 00:08:45.439 }, 00:08:45.439 { 00:08:45.439 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:45.439 "dma_device_type": 2 00:08:45.439 } 00:08:45.439 ], 00:08:45.439 "driver_specific": {} 00:08:45.439 } 00:08:45.439 ] 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.439 "name": "Existed_Raid", 00:08:45.439 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:45.439 "strip_size_kb": 0, 00:08:45.439 "state": "configuring", 00:08:45.439 "raid_level": "raid1", 00:08:45.439 "superblock": false, 00:08:45.439 "num_base_bdevs": 3, 00:08:45.439 "num_base_bdevs_discovered": 2, 00:08:45.439 "num_base_bdevs_operational": 3, 00:08:45.439 "base_bdevs_list": [ 00:08:45.439 { 00:08:45.439 "name": "BaseBdev1", 00:08:45.439 "uuid": "64bd938c-32e3-485d-a66f-4fd459c0db01", 00:08:45.439 "is_configured": true, 00:08:45.439 "data_offset": 0, 00:08:45.439 "data_size": 65536 00:08:45.439 }, 00:08:45.439 { 00:08:45.439 "name": "BaseBdev2", 00:08:45.439 "uuid": "ea9f4306-89c3-4043-9763-a2cfa076d3c8", 00:08:45.439 "is_configured": true, 00:08:45.439 "data_offset": 0, 00:08:45.439 "data_size": 65536 00:08:45.439 }, 00:08:45.439 { 00:08:45.439 "name": "BaseBdev3", 00:08:45.439 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:45.439 "is_configured": false, 00:08:45.439 "data_offset": 0, 00:08:45.439 "data_size": 0 00:08:45.439 } 00:08:45.439 ] 00:08:45.439 }' 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.439 05:36:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.007 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:46.007 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.007 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.007 [2024-12-07 05:36:19.094768] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:46.008 [2024-12-07 05:36:19.094831] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:46.008 [2024-12-07 05:36:19.094859] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:46.008 [2024-12-07 05:36:19.095256] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:46.008 [2024-12-07 05:36:19.095462] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:46.008 [2024-12-07 05:36:19.095487] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:46.008 [2024-12-07 05:36:19.095781] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:46.008 BaseBdev3 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.008 [ 00:08:46.008 { 00:08:46.008 "name": "BaseBdev3", 00:08:46.008 "aliases": [ 00:08:46.008 "43108fd2-e67e-4e0e-ab69-e720c49df816" 00:08:46.008 ], 00:08:46.008 "product_name": "Malloc disk", 00:08:46.008 "block_size": 512, 00:08:46.008 "num_blocks": 65536, 00:08:46.008 "uuid": "43108fd2-e67e-4e0e-ab69-e720c49df816", 00:08:46.008 "assigned_rate_limits": { 00:08:46.008 "rw_ios_per_sec": 0, 00:08:46.008 "rw_mbytes_per_sec": 0, 00:08:46.008 "r_mbytes_per_sec": 0, 00:08:46.008 "w_mbytes_per_sec": 0 00:08:46.008 }, 00:08:46.008 "claimed": true, 00:08:46.008 "claim_type": "exclusive_write", 00:08:46.008 "zoned": false, 00:08:46.008 "supported_io_types": { 00:08:46.008 "read": true, 00:08:46.008 "write": true, 00:08:46.008 "unmap": true, 00:08:46.008 "flush": true, 00:08:46.008 "reset": true, 00:08:46.008 "nvme_admin": false, 00:08:46.008 "nvme_io": false, 00:08:46.008 "nvme_io_md": false, 00:08:46.008 "write_zeroes": true, 00:08:46.008 "zcopy": true, 00:08:46.008 "get_zone_info": false, 00:08:46.008 "zone_management": false, 00:08:46.008 "zone_append": false, 00:08:46.008 "compare": false, 00:08:46.008 "compare_and_write": false, 00:08:46.008 "abort": true, 00:08:46.008 "seek_hole": false, 00:08:46.008 "seek_data": false, 00:08:46.008 "copy": true, 00:08:46.008 "nvme_iov_md": false 00:08:46.008 }, 00:08:46.008 "memory_domains": [ 00:08:46.008 { 00:08:46.008 "dma_device_id": "system", 00:08:46.008 "dma_device_type": 1 00:08:46.008 }, 00:08:46.008 { 00:08:46.008 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.008 "dma_device_type": 2 00:08:46.008 } 00:08:46.008 ], 00:08:46.008 "driver_specific": {} 00:08:46.008 } 00:08:46.008 ] 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:46.008 "name": "Existed_Raid", 00:08:46.008 "uuid": "bd92a1ab-d6d3-4268-9129-442eedb27e56", 00:08:46.008 "strip_size_kb": 0, 00:08:46.008 "state": "online", 00:08:46.008 "raid_level": "raid1", 00:08:46.008 "superblock": false, 00:08:46.008 "num_base_bdevs": 3, 00:08:46.008 "num_base_bdevs_discovered": 3, 00:08:46.008 "num_base_bdevs_operational": 3, 00:08:46.008 "base_bdevs_list": [ 00:08:46.008 { 00:08:46.008 "name": "BaseBdev1", 00:08:46.008 "uuid": "64bd938c-32e3-485d-a66f-4fd459c0db01", 00:08:46.008 "is_configured": true, 00:08:46.008 "data_offset": 0, 00:08:46.008 "data_size": 65536 00:08:46.008 }, 00:08:46.008 { 00:08:46.008 "name": "BaseBdev2", 00:08:46.008 "uuid": "ea9f4306-89c3-4043-9763-a2cfa076d3c8", 00:08:46.008 "is_configured": true, 00:08:46.008 "data_offset": 0, 00:08:46.008 "data_size": 65536 00:08:46.008 }, 00:08:46.008 { 00:08:46.008 "name": "BaseBdev3", 00:08:46.008 "uuid": "43108fd2-e67e-4e0e-ab69-e720c49df816", 00:08:46.008 "is_configured": true, 00:08:46.008 "data_offset": 0, 00:08:46.008 "data_size": 65536 00:08:46.008 } 00:08:46.008 ] 00:08:46.008 }' 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:46.008 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.267 [2024-12-07 05:36:19.558342] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:46.267 "name": "Existed_Raid", 00:08:46.267 "aliases": [ 00:08:46.267 "bd92a1ab-d6d3-4268-9129-442eedb27e56" 00:08:46.267 ], 00:08:46.267 "product_name": "Raid Volume", 00:08:46.267 "block_size": 512, 00:08:46.267 "num_blocks": 65536, 00:08:46.267 "uuid": "bd92a1ab-d6d3-4268-9129-442eedb27e56", 00:08:46.267 "assigned_rate_limits": { 00:08:46.267 "rw_ios_per_sec": 0, 00:08:46.267 "rw_mbytes_per_sec": 0, 00:08:46.267 "r_mbytes_per_sec": 0, 00:08:46.267 "w_mbytes_per_sec": 0 00:08:46.267 }, 00:08:46.267 "claimed": false, 00:08:46.267 "zoned": false, 00:08:46.267 "supported_io_types": { 00:08:46.267 "read": true, 00:08:46.267 "write": true, 00:08:46.267 "unmap": false, 00:08:46.267 "flush": false, 00:08:46.267 "reset": true, 00:08:46.267 "nvme_admin": false, 00:08:46.267 "nvme_io": false, 00:08:46.267 "nvme_io_md": false, 00:08:46.267 "write_zeroes": true, 00:08:46.267 "zcopy": false, 00:08:46.267 "get_zone_info": false, 00:08:46.267 "zone_management": false, 00:08:46.267 "zone_append": false, 00:08:46.267 "compare": false, 00:08:46.267 "compare_and_write": false, 00:08:46.267 "abort": false, 00:08:46.267 "seek_hole": false, 00:08:46.267 "seek_data": false, 00:08:46.267 "copy": false, 00:08:46.267 "nvme_iov_md": false 00:08:46.267 }, 00:08:46.267 "memory_domains": [ 00:08:46.267 { 00:08:46.267 "dma_device_id": "system", 00:08:46.267 "dma_device_type": 1 00:08:46.267 }, 00:08:46.267 { 00:08:46.267 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.267 "dma_device_type": 2 00:08:46.267 }, 00:08:46.267 { 00:08:46.267 "dma_device_id": "system", 00:08:46.267 "dma_device_type": 1 00:08:46.267 }, 00:08:46.267 { 00:08:46.267 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.267 "dma_device_type": 2 00:08:46.267 }, 00:08:46.267 { 00:08:46.267 "dma_device_id": "system", 00:08:46.267 "dma_device_type": 1 00:08:46.267 }, 00:08:46.267 { 00:08:46.267 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.267 "dma_device_type": 2 00:08:46.267 } 00:08:46.267 ], 00:08:46.267 "driver_specific": { 00:08:46.267 "raid": { 00:08:46.267 "uuid": "bd92a1ab-d6d3-4268-9129-442eedb27e56", 00:08:46.267 "strip_size_kb": 0, 00:08:46.267 "state": "online", 00:08:46.267 "raid_level": "raid1", 00:08:46.267 "superblock": false, 00:08:46.267 "num_base_bdevs": 3, 00:08:46.267 "num_base_bdevs_discovered": 3, 00:08:46.267 "num_base_bdevs_operational": 3, 00:08:46.267 "base_bdevs_list": [ 00:08:46.267 { 00:08:46.267 "name": "BaseBdev1", 00:08:46.267 "uuid": "64bd938c-32e3-485d-a66f-4fd459c0db01", 00:08:46.267 "is_configured": true, 00:08:46.267 "data_offset": 0, 00:08:46.267 "data_size": 65536 00:08:46.267 }, 00:08:46.267 { 00:08:46.267 "name": "BaseBdev2", 00:08:46.267 "uuid": "ea9f4306-89c3-4043-9763-a2cfa076d3c8", 00:08:46.267 "is_configured": true, 00:08:46.267 "data_offset": 0, 00:08:46.267 "data_size": 65536 00:08:46.267 }, 00:08:46.267 { 00:08:46.267 "name": "BaseBdev3", 00:08:46.267 "uuid": "43108fd2-e67e-4e0e-ab69-e720c49df816", 00:08:46.267 "is_configured": true, 00:08:46.267 "data_offset": 0, 00:08:46.267 "data_size": 65536 00:08:46.267 } 00:08:46.267 ] 00:08:46.267 } 00:08:46.267 } 00:08:46.267 }' 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:46.267 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:46.268 BaseBdev2 00:08:46.268 BaseBdev3' 00:08:46.268 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.526 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:46.526 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:46.526 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.527 [2024-12-07 05:36:19.789647] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:46.527 "name": "Existed_Raid", 00:08:46.527 "uuid": "bd92a1ab-d6d3-4268-9129-442eedb27e56", 00:08:46.527 "strip_size_kb": 0, 00:08:46.527 "state": "online", 00:08:46.527 "raid_level": "raid1", 00:08:46.527 "superblock": false, 00:08:46.527 "num_base_bdevs": 3, 00:08:46.527 "num_base_bdevs_discovered": 2, 00:08:46.527 "num_base_bdevs_operational": 2, 00:08:46.527 "base_bdevs_list": [ 00:08:46.527 { 00:08:46.527 "name": null, 00:08:46.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:46.527 "is_configured": false, 00:08:46.527 "data_offset": 0, 00:08:46.527 "data_size": 65536 00:08:46.527 }, 00:08:46.527 { 00:08:46.527 "name": "BaseBdev2", 00:08:46.527 "uuid": "ea9f4306-89c3-4043-9763-a2cfa076d3c8", 00:08:46.527 "is_configured": true, 00:08:46.527 "data_offset": 0, 00:08:46.527 "data_size": 65536 00:08:46.527 }, 00:08:46.527 { 00:08:46.527 "name": "BaseBdev3", 00:08:46.527 "uuid": "43108fd2-e67e-4e0e-ab69-e720c49df816", 00:08:46.527 "is_configured": true, 00:08:46.527 "data_offset": 0, 00:08:46.527 "data_size": 65536 00:08:46.527 } 00:08:46.527 ] 00:08:46.527 }' 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:46.527 05:36:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 [2024-12-07 05:36:20.272254] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 [2024-12-07 05:36:20.323610] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:47.096 [2024-12-07 05:36:20.323720] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:47.096 [2024-12-07 05:36:20.335565] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:47.096 [2024-12-07 05:36:20.335620] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:47.096 [2024-12-07 05:36:20.335646] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 BaseBdev2 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 [ 00:08:47.096 { 00:08:47.096 "name": "BaseBdev2", 00:08:47.096 "aliases": [ 00:08:47.096 "16ffb42b-93f5-4fe1-bbaa-d04b8ba30a0a" 00:08:47.096 ], 00:08:47.096 "product_name": "Malloc disk", 00:08:47.096 "block_size": 512, 00:08:47.096 "num_blocks": 65536, 00:08:47.096 "uuid": "16ffb42b-93f5-4fe1-bbaa-d04b8ba30a0a", 00:08:47.096 "assigned_rate_limits": { 00:08:47.096 "rw_ios_per_sec": 0, 00:08:47.096 "rw_mbytes_per_sec": 0, 00:08:47.096 "r_mbytes_per_sec": 0, 00:08:47.096 "w_mbytes_per_sec": 0 00:08:47.096 }, 00:08:47.096 "claimed": false, 00:08:47.096 "zoned": false, 00:08:47.096 "supported_io_types": { 00:08:47.096 "read": true, 00:08:47.096 "write": true, 00:08:47.096 "unmap": true, 00:08:47.096 "flush": true, 00:08:47.096 "reset": true, 00:08:47.096 "nvme_admin": false, 00:08:47.096 "nvme_io": false, 00:08:47.096 "nvme_io_md": false, 00:08:47.096 "write_zeroes": true, 00:08:47.096 "zcopy": true, 00:08:47.096 "get_zone_info": false, 00:08:47.096 "zone_management": false, 00:08:47.096 "zone_append": false, 00:08:47.096 "compare": false, 00:08:47.096 "compare_and_write": false, 00:08:47.096 "abort": true, 00:08:47.096 "seek_hole": false, 00:08:47.096 "seek_data": false, 00:08:47.096 "copy": true, 00:08:47.096 "nvme_iov_md": false 00:08:47.096 }, 00:08:47.096 "memory_domains": [ 00:08:47.096 { 00:08:47.096 "dma_device_id": "system", 00:08:47.096 "dma_device_type": 1 00:08:47.096 }, 00:08:47.096 { 00:08:47.096 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:47.096 "dma_device_type": 2 00:08:47.096 } 00:08:47.096 ], 00:08:47.096 "driver_specific": {} 00:08:47.096 } 00:08:47.096 ] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 BaseBdev3 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.096 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.096 [ 00:08:47.097 { 00:08:47.097 "name": "BaseBdev3", 00:08:47.097 "aliases": [ 00:08:47.097 "d6126cd3-44af-4e7a-843c-5ac08b9ba259" 00:08:47.097 ], 00:08:47.097 "product_name": "Malloc disk", 00:08:47.097 "block_size": 512, 00:08:47.097 "num_blocks": 65536, 00:08:47.097 "uuid": "d6126cd3-44af-4e7a-843c-5ac08b9ba259", 00:08:47.097 "assigned_rate_limits": { 00:08:47.097 "rw_ios_per_sec": 0, 00:08:47.097 "rw_mbytes_per_sec": 0, 00:08:47.097 "r_mbytes_per_sec": 0, 00:08:47.097 "w_mbytes_per_sec": 0 00:08:47.097 }, 00:08:47.097 "claimed": false, 00:08:47.097 "zoned": false, 00:08:47.097 "supported_io_types": { 00:08:47.097 "read": true, 00:08:47.097 "write": true, 00:08:47.097 "unmap": true, 00:08:47.097 "flush": true, 00:08:47.097 "reset": true, 00:08:47.097 "nvme_admin": false, 00:08:47.097 "nvme_io": false, 00:08:47.097 "nvme_io_md": false, 00:08:47.097 "write_zeroes": true, 00:08:47.097 "zcopy": true, 00:08:47.097 "get_zone_info": false, 00:08:47.097 "zone_management": false, 00:08:47.097 "zone_append": false, 00:08:47.097 "compare": false, 00:08:47.097 "compare_and_write": false, 00:08:47.097 "abort": true, 00:08:47.097 "seek_hole": false, 00:08:47.097 "seek_data": false, 00:08:47.097 "copy": true, 00:08:47.097 "nvme_iov_md": false 00:08:47.097 }, 00:08:47.097 "memory_domains": [ 00:08:47.097 { 00:08:47.097 "dma_device_id": "system", 00:08:47.097 "dma_device_type": 1 00:08:47.097 }, 00:08:47.097 { 00:08:47.097 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:47.097 "dma_device_type": 2 00:08:47.097 } 00:08:47.097 ], 00:08:47.097 "driver_specific": {} 00:08:47.097 } 00:08:47.097 ] 00:08:47.097 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.097 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:47.097 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:47.097 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:47.097 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:47.097 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.097 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.097 [2024-12-07 05:36:20.460283] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:47.097 [2024-12-07 05:36:20.460330] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:47.097 [2024-12-07 05:36:20.460354] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:47.356 [2024-12-07 05:36:20.462235] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:47.356 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.356 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:47.356 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:47.356 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:47.356 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:47.356 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:47.356 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:47.357 "name": "Existed_Raid", 00:08:47.357 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.357 "strip_size_kb": 0, 00:08:47.357 "state": "configuring", 00:08:47.357 "raid_level": "raid1", 00:08:47.357 "superblock": false, 00:08:47.357 "num_base_bdevs": 3, 00:08:47.357 "num_base_bdevs_discovered": 2, 00:08:47.357 "num_base_bdevs_operational": 3, 00:08:47.357 "base_bdevs_list": [ 00:08:47.357 { 00:08:47.357 "name": "BaseBdev1", 00:08:47.357 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.357 "is_configured": false, 00:08:47.357 "data_offset": 0, 00:08:47.357 "data_size": 0 00:08:47.357 }, 00:08:47.357 { 00:08:47.357 "name": "BaseBdev2", 00:08:47.357 "uuid": "16ffb42b-93f5-4fe1-bbaa-d04b8ba30a0a", 00:08:47.357 "is_configured": true, 00:08:47.357 "data_offset": 0, 00:08:47.357 "data_size": 65536 00:08:47.357 }, 00:08:47.357 { 00:08:47.357 "name": "BaseBdev3", 00:08:47.357 "uuid": "d6126cd3-44af-4e7a-843c-5ac08b9ba259", 00:08:47.357 "is_configured": true, 00:08:47.357 "data_offset": 0, 00:08:47.357 "data_size": 65536 00:08:47.357 } 00:08:47.357 ] 00:08:47.357 }' 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:47.357 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.616 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:47.616 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.616 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.616 [2024-12-07 05:36:20.871585] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:47.616 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.616 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:47.616 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:47.617 "name": "Existed_Raid", 00:08:47.617 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.617 "strip_size_kb": 0, 00:08:47.617 "state": "configuring", 00:08:47.617 "raid_level": "raid1", 00:08:47.617 "superblock": false, 00:08:47.617 "num_base_bdevs": 3, 00:08:47.617 "num_base_bdevs_discovered": 1, 00:08:47.617 "num_base_bdevs_operational": 3, 00:08:47.617 "base_bdevs_list": [ 00:08:47.617 { 00:08:47.617 "name": "BaseBdev1", 00:08:47.617 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:47.617 "is_configured": false, 00:08:47.617 "data_offset": 0, 00:08:47.617 "data_size": 0 00:08:47.617 }, 00:08:47.617 { 00:08:47.617 "name": null, 00:08:47.617 "uuid": "16ffb42b-93f5-4fe1-bbaa-d04b8ba30a0a", 00:08:47.617 "is_configured": false, 00:08:47.617 "data_offset": 0, 00:08:47.617 "data_size": 65536 00:08:47.617 }, 00:08:47.617 { 00:08:47.617 "name": "BaseBdev3", 00:08:47.617 "uuid": "d6126cd3-44af-4e7a-843c-5ac08b9ba259", 00:08:47.617 "is_configured": true, 00:08:47.617 "data_offset": 0, 00:08:47.617 "data_size": 65536 00:08:47.617 } 00:08:47.617 ] 00:08:47.617 }' 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:47.617 05:36:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.186 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.186 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:48.186 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.186 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.186 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.186 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:48.186 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.187 [2024-12-07 05:36:21.385672] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:48.187 BaseBdev1 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.187 [ 00:08:48.187 { 00:08:48.187 "name": "BaseBdev1", 00:08:48.187 "aliases": [ 00:08:48.187 "e1bad4a4-c9e5-4967-927e-0e440b5147ef" 00:08:48.187 ], 00:08:48.187 "product_name": "Malloc disk", 00:08:48.187 "block_size": 512, 00:08:48.187 "num_blocks": 65536, 00:08:48.187 "uuid": "e1bad4a4-c9e5-4967-927e-0e440b5147ef", 00:08:48.187 "assigned_rate_limits": { 00:08:48.187 "rw_ios_per_sec": 0, 00:08:48.187 "rw_mbytes_per_sec": 0, 00:08:48.187 "r_mbytes_per_sec": 0, 00:08:48.187 "w_mbytes_per_sec": 0 00:08:48.187 }, 00:08:48.187 "claimed": true, 00:08:48.187 "claim_type": "exclusive_write", 00:08:48.187 "zoned": false, 00:08:48.187 "supported_io_types": { 00:08:48.187 "read": true, 00:08:48.187 "write": true, 00:08:48.187 "unmap": true, 00:08:48.187 "flush": true, 00:08:48.187 "reset": true, 00:08:48.187 "nvme_admin": false, 00:08:48.187 "nvme_io": false, 00:08:48.187 "nvme_io_md": false, 00:08:48.187 "write_zeroes": true, 00:08:48.187 "zcopy": true, 00:08:48.187 "get_zone_info": false, 00:08:48.187 "zone_management": false, 00:08:48.187 "zone_append": false, 00:08:48.187 "compare": false, 00:08:48.187 "compare_and_write": false, 00:08:48.187 "abort": true, 00:08:48.187 "seek_hole": false, 00:08:48.187 "seek_data": false, 00:08:48.187 "copy": true, 00:08:48.187 "nvme_iov_md": false 00:08:48.187 }, 00:08:48.187 "memory_domains": [ 00:08:48.187 { 00:08:48.187 "dma_device_id": "system", 00:08:48.187 "dma_device_type": 1 00:08:48.187 }, 00:08:48.187 { 00:08:48.187 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.187 "dma_device_type": 2 00:08:48.187 } 00:08:48.187 ], 00:08:48.187 "driver_specific": {} 00:08:48.187 } 00:08:48.187 ] 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:48.187 "name": "Existed_Raid", 00:08:48.187 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.187 "strip_size_kb": 0, 00:08:48.187 "state": "configuring", 00:08:48.187 "raid_level": "raid1", 00:08:48.187 "superblock": false, 00:08:48.187 "num_base_bdevs": 3, 00:08:48.187 "num_base_bdevs_discovered": 2, 00:08:48.187 "num_base_bdevs_operational": 3, 00:08:48.187 "base_bdevs_list": [ 00:08:48.187 { 00:08:48.187 "name": "BaseBdev1", 00:08:48.187 "uuid": "e1bad4a4-c9e5-4967-927e-0e440b5147ef", 00:08:48.187 "is_configured": true, 00:08:48.187 "data_offset": 0, 00:08:48.187 "data_size": 65536 00:08:48.187 }, 00:08:48.187 { 00:08:48.187 "name": null, 00:08:48.187 "uuid": "16ffb42b-93f5-4fe1-bbaa-d04b8ba30a0a", 00:08:48.187 "is_configured": false, 00:08:48.187 "data_offset": 0, 00:08:48.187 "data_size": 65536 00:08:48.187 }, 00:08:48.187 { 00:08:48.187 "name": "BaseBdev3", 00:08:48.187 "uuid": "d6126cd3-44af-4e7a-843c-5ac08b9ba259", 00:08:48.187 "is_configured": true, 00:08:48.187 "data_offset": 0, 00:08:48.187 "data_size": 65536 00:08:48.187 } 00:08:48.187 ] 00:08:48.187 }' 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:48.187 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.756 [2024-12-07 05:36:21.924791] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:48.756 "name": "Existed_Raid", 00:08:48.756 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.756 "strip_size_kb": 0, 00:08:48.756 "state": "configuring", 00:08:48.756 "raid_level": "raid1", 00:08:48.756 "superblock": false, 00:08:48.756 "num_base_bdevs": 3, 00:08:48.756 "num_base_bdevs_discovered": 1, 00:08:48.756 "num_base_bdevs_operational": 3, 00:08:48.756 "base_bdevs_list": [ 00:08:48.756 { 00:08:48.756 "name": "BaseBdev1", 00:08:48.756 "uuid": "e1bad4a4-c9e5-4967-927e-0e440b5147ef", 00:08:48.756 "is_configured": true, 00:08:48.756 "data_offset": 0, 00:08:48.756 "data_size": 65536 00:08:48.756 }, 00:08:48.756 { 00:08:48.756 "name": null, 00:08:48.756 "uuid": "16ffb42b-93f5-4fe1-bbaa-d04b8ba30a0a", 00:08:48.756 "is_configured": false, 00:08:48.756 "data_offset": 0, 00:08:48.756 "data_size": 65536 00:08:48.756 }, 00:08:48.756 { 00:08:48.756 "name": null, 00:08:48.756 "uuid": "d6126cd3-44af-4e7a-843c-5ac08b9ba259", 00:08:48.756 "is_configured": false, 00:08:48.756 "data_offset": 0, 00:08:48.756 "data_size": 65536 00:08:48.756 } 00:08:48.756 ] 00:08:48.756 }' 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:48.756 05:36:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.016 [2024-12-07 05:36:22.364115] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.016 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.275 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.275 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.275 "name": "Existed_Raid", 00:08:49.275 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:49.275 "strip_size_kb": 0, 00:08:49.275 "state": "configuring", 00:08:49.275 "raid_level": "raid1", 00:08:49.275 "superblock": false, 00:08:49.275 "num_base_bdevs": 3, 00:08:49.275 "num_base_bdevs_discovered": 2, 00:08:49.275 "num_base_bdevs_operational": 3, 00:08:49.275 "base_bdevs_list": [ 00:08:49.275 { 00:08:49.275 "name": "BaseBdev1", 00:08:49.275 "uuid": "e1bad4a4-c9e5-4967-927e-0e440b5147ef", 00:08:49.275 "is_configured": true, 00:08:49.275 "data_offset": 0, 00:08:49.275 "data_size": 65536 00:08:49.275 }, 00:08:49.275 { 00:08:49.275 "name": null, 00:08:49.275 "uuid": "16ffb42b-93f5-4fe1-bbaa-d04b8ba30a0a", 00:08:49.275 "is_configured": false, 00:08:49.275 "data_offset": 0, 00:08:49.275 "data_size": 65536 00:08:49.275 }, 00:08:49.275 { 00:08:49.275 "name": "BaseBdev3", 00:08:49.275 "uuid": "d6126cd3-44af-4e7a-843c-5ac08b9ba259", 00:08:49.275 "is_configured": true, 00:08:49.275 "data_offset": 0, 00:08:49.275 "data_size": 65536 00:08:49.275 } 00:08:49.275 ] 00:08:49.275 }' 00:08:49.275 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.275 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.536 [2024-12-07 05:36:22.803401] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.536 "name": "Existed_Raid", 00:08:49.536 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:49.536 "strip_size_kb": 0, 00:08:49.536 "state": "configuring", 00:08:49.536 "raid_level": "raid1", 00:08:49.536 "superblock": false, 00:08:49.536 "num_base_bdevs": 3, 00:08:49.536 "num_base_bdevs_discovered": 1, 00:08:49.536 "num_base_bdevs_operational": 3, 00:08:49.536 "base_bdevs_list": [ 00:08:49.536 { 00:08:49.536 "name": null, 00:08:49.536 "uuid": "e1bad4a4-c9e5-4967-927e-0e440b5147ef", 00:08:49.536 "is_configured": false, 00:08:49.536 "data_offset": 0, 00:08:49.536 "data_size": 65536 00:08:49.536 }, 00:08:49.536 { 00:08:49.536 "name": null, 00:08:49.536 "uuid": "16ffb42b-93f5-4fe1-bbaa-d04b8ba30a0a", 00:08:49.536 "is_configured": false, 00:08:49.536 "data_offset": 0, 00:08:49.536 "data_size": 65536 00:08:49.536 }, 00:08:49.536 { 00:08:49.536 "name": "BaseBdev3", 00:08:49.536 "uuid": "d6126cd3-44af-4e7a-843c-5ac08b9ba259", 00:08:49.536 "is_configured": true, 00:08:49.536 "data_offset": 0, 00:08:49.536 "data_size": 65536 00:08:49.536 } 00:08:49.536 ] 00:08:49.536 }' 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.536 05:36:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 [2024-12-07 05:36:23.261233] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.170 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:50.170 "name": "Existed_Raid", 00:08:50.170 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:50.170 "strip_size_kb": 0, 00:08:50.170 "state": "configuring", 00:08:50.170 "raid_level": "raid1", 00:08:50.170 "superblock": false, 00:08:50.170 "num_base_bdevs": 3, 00:08:50.171 "num_base_bdevs_discovered": 2, 00:08:50.171 "num_base_bdevs_operational": 3, 00:08:50.171 "base_bdevs_list": [ 00:08:50.171 { 00:08:50.171 "name": null, 00:08:50.171 "uuid": "e1bad4a4-c9e5-4967-927e-0e440b5147ef", 00:08:50.171 "is_configured": false, 00:08:50.171 "data_offset": 0, 00:08:50.171 "data_size": 65536 00:08:50.171 }, 00:08:50.171 { 00:08:50.171 "name": "BaseBdev2", 00:08:50.171 "uuid": "16ffb42b-93f5-4fe1-bbaa-d04b8ba30a0a", 00:08:50.171 "is_configured": true, 00:08:50.171 "data_offset": 0, 00:08:50.171 "data_size": 65536 00:08:50.171 }, 00:08:50.171 { 00:08:50.171 "name": "BaseBdev3", 00:08:50.171 "uuid": "d6126cd3-44af-4e7a-843c-5ac08b9ba259", 00:08:50.171 "is_configured": true, 00:08:50.171 "data_offset": 0, 00:08:50.171 "data_size": 65536 00:08:50.171 } 00:08:50.171 ] 00:08:50.171 }' 00:08:50.171 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:50.171 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u e1bad4a4-c9e5-4967-927e-0e440b5147ef 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.430 [2024-12-07 05:36:23.779248] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:50.430 [2024-12-07 05:36:23.779368] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:50.430 [2024-12-07 05:36:23.779396] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:50.430 [2024-12-07 05:36:23.779695] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:50.430 [2024-12-07 05:36:23.779859] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:50.430 [2024-12-07 05:36:23.779903] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:50.430 [2024-12-07 05:36:23.780120] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:50.430 NewBaseBdev 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.430 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.690 [ 00:08:50.690 { 00:08:50.690 "name": "NewBaseBdev", 00:08:50.690 "aliases": [ 00:08:50.690 "e1bad4a4-c9e5-4967-927e-0e440b5147ef" 00:08:50.690 ], 00:08:50.691 "product_name": "Malloc disk", 00:08:50.691 "block_size": 512, 00:08:50.691 "num_blocks": 65536, 00:08:50.691 "uuid": "e1bad4a4-c9e5-4967-927e-0e440b5147ef", 00:08:50.691 "assigned_rate_limits": { 00:08:50.691 "rw_ios_per_sec": 0, 00:08:50.691 "rw_mbytes_per_sec": 0, 00:08:50.691 "r_mbytes_per_sec": 0, 00:08:50.691 "w_mbytes_per_sec": 0 00:08:50.691 }, 00:08:50.691 "claimed": true, 00:08:50.691 "claim_type": "exclusive_write", 00:08:50.691 "zoned": false, 00:08:50.691 "supported_io_types": { 00:08:50.691 "read": true, 00:08:50.691 "write": true, 00:08:50.691 "unmap": true, 00:08:50.691 "flush": true, 00:08:50.691 "reset": true, 00:08:50.691 "nvme_admin": false, 00:08:50.691 "nvme_io": false, 00:08:50.691 "nvme_io_md": false, 00:08:50.691 "write_zeroes": true, 00:08:50.691 "zcopy": true, 00:08:50.691 "get_zone_info": false, 00:08:50.691 "zone_management": false, 00:08:50.691 "zone_append": false, 00:08:50.691 "compare": false, 00:08:50.691 "compare_and_write": false, 00:08:50.691 "abort": true, 00:08:50.691 "seek_hole": false, 00:08:50.691 "seek_data": false, 00:08:50.691 "copy": true, 00:08:50.691 "nvme_iov_md": false 00:08:50.691 }, 00:08:50.691 "memory_domains": [ 00:08:50.691 { 00:08:50.691 "dma_device_id": "system", 00:08:50.691 "dma_device_type": 1 00:08:50.691 }, 00:08:50.691 { 00:08:50.691 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.691 "dma_device_type": 2 00:08:50.691 } 00:08:50.691 ], 00:08:50.691 "driver_specific": {} 00:08:50.691 } 00:08:50.691 ] 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:50.691 "name": "Existed_Raid", 00:08:50.691 "uuid": "b8a9e02b-864c-4dcc-8ee0-cf353a67ce94", 00:08:50.691 "strip_size_kb": 0, 00:08:50.691 "state": "online", 00:08:50.691 "raid_level": "raid1", 00:08:50.691 "superblock": false, 00:08:50.691 "num_base_bdevs": 3, 00:08:50.691 "num_base_bdevs_discovered": 3, 00:08:50.691 "num_base_bdevs_operational": 3, 00:08:50.691 "base_bdevs_list": [ 00:08:50.691 { 00:08:50.691 "name": "NewBaseBdev", 00:08:50.691 "uuid": "e1bad4a4-c9e5-4967-927e-0e440b5147ef", 00:08:50.691 "is_configured": true, 00:08:50.691 "data_offset": 0, 00:08:50.691 "data_size": 65536 00:08:50.691 }, 00:08:50.691 { 00:08:50.691 "name": "BaseBdev2", 00:08:50.691 "uuid": "16ffb42b-93f5-4fe1-bbaa-d04b8ba30a0a", 00:08:50.691 "is_configured": true, 00:08:50.691 "data_offset": 0, 00:08:50.691 "data_size": 65536 00:08:50.691 }, 00:08:50.691 { 00:08:50.691 "name": "BaseBdev3", 00:08:50.691 "uuid": "d6126cd3-44af-4e7a-843c-5ac08b9ba259", 00:08:50.691 "is_configured": true, 00:08:50.691 "data_offset": 0, 00:08:50.691 "data_size": 65536 00:08:50.691 } 00:08:50.691 ] 00:08:50.691 }' 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:50.691 05:36:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.950 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:50.950 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:50.950 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:50.950 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:50.950 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:50.950 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:50.950 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:50.950 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:50.950 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.950 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.951 [2024-12-07 05:36:24.214954] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:50.951 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.951 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:50.951 "name": "Existed_Raid", 00:08:50.951 "aliases": [ 00:08:50.951 "b8a9e02b-864c-4dcc-8ee0-cf353a67ce94" 00:08:50.951 ], 00:08:50.951 "product_name": "Raid Volume", 00:08:50.951 "block_size": 512, 00:08:50.951 "num_blocks": 65536, 00:08:50.951 "uuid": "b8a9e02b-864c-4dcc-8ee0-cf353a67ce94", 00:08:50.951 "assigned_rate_limits": { 00:08:50.951 "rw_ios_per_sec": 0, 00:08:50.951 "rw_mbytes_per_sec": 0, 00:08:50.951 "r_mbytes_per_sec": 0, 00:08:50.951 "w_mbytes_per_sec": 0 00:08:50.951 }, 00:08:50.951 "claimed": false, 00:08:50.951 "zoned": false, 00:08:50.951 "supported_io_types": { 00:08:50.951 "read": true, 00:08:50.951 "write": true, 00:08:50.951 "unmap": false, 00:08:50.951 "flush": false, 00:08:50.951 "reset": true, 00:08:50.951 "nvme_admin": false, 00:08:50.951 "nvme_io": false, 00:08:50.951 "nvme_io_md": false, 00:08:50.951 "write_zeroes": true, 00:08:50.951 "zcopy": false, 00:08:50.951 "get_zone_info": false, 00:08:50.951 "zone_management": false, 00:08:50.951 "zone_append": false, 00:08:50.951 "compare": false, 00:08:50.951 "compare_and_write": false, 00:08:50.951 "abort": false, 00:08:50.951 "seek_hole": false, 00:08:50.951 "seek_data": false, 00:08:50.951 "copy": false, 00:08:50.951 "nvme_iov_md": false 00:08:50.951 }, 00:08:50.951 "memory_domains": [ 00:08:50.951 { 00:08:50.951 "dma_device_id": "system", 00:08:50.951 "dma_device_type": 1 00:08:50.951 }, 00:08:50.951 { 00:08:50.951 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.951 "dma_device_type": 2 00:08:50.951 }, 00:08:50.951 { 00:08:50.951 "dma_device_id": "system", 00:08:50.951 "dma_device_type": 1 00:08:50.951 }, 00:08:50.951 { 00:08:50.951 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.951 "dma_device_type": 2 00:08:50.951 }, 00:08:50.951 { 00:08:50.951 "dma_device_id": "system", 00:08:50.951 "dma_device_type": 1 00:08:50.951 }, 00:08:50.951 { 00:08:50.951 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.951 "dma_device_type": 2 00:08:50.951 } 00:08:50.951 ], 00:08:50.951 "driver_specific": { 00:08:50.951 "raid": { 00:08:50.951 "uuid": "b8a9e02b-864c-4dcc-8ee0-cf353a67ce94", 00:08:50.951 "strip_size_kb": 0, 00:08:50.951 "state": "online", 00:08:50.951 "raid_level": "raid1", 00:08:50.951 "superblock": false, 00:08:50.951 "num_base_bdevs": 3, 00:08:50.951 "num_base_bdevs_discovered": 3, 00:08:50.951 "num_base_bdevs_operational": 3, 00:08:50.951 "base_bdevs_list": [ 00:08:50.951 { 00:08:50.951 "name": "NewBaseBdev", 00:08:50.951 "uuid": "e1bad4a4-c9e5-4967-927e-0e440b5147ef", 00:08:50.951 "is_configured": true, 00:08:50.951 "data_offset": 0, 00:08:50.951 "data_size": 65536 00:08:50.951 }, 00:08:50.951 { 00:08:50.951 "name": "BaseBdev2", 00:08:50.951 "uuid": "16ffb42b-93f5-4fe1-bbaa-d04b8ba30a0a", 00:08:50.951 "is_configured": true, 00:08:50.951 "data_offset": 0, 00:08:50.951 "data_size": 65536 00:08:50.951 }, 00:08:50.951 { 00:08:50.951 "name": "BaseBdev3", 00:08:50.951 "uuid": "d6126cd3-44af-4e7a-843c-5ac08b9ba259", 00:08:50.951 "is_configured": true, 00:08:50.951 "data_offset": 0, 00:08:50.951 "data_size": 65536 00:08:50.951 } 00:08:50.951 ] 00:08:50.951 } 00:08:50.951 } 00:08:50.951 }' 00:08:50.951 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:50.951 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:50.951 BaseBdev2 00:08:50.951 BaseBdev3' 00:08:50.951 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:51.209 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:51.209 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:51.209 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:51.209 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:51.209 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.209 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.210 [2024-12-07 05:36:24.506141] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:51.210 [2024-12-07 05:36:24.506169] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:51.210 [2024-12-07 05:36:24.506238] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:51.210 [2024-12-07 05:36:24.506510] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:51.210 [2024-12-07 05:36:24.506521] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 78115 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 78115 ']' 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 78115 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78115 00:08:51.210 killing process with pid 78115 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78115' 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 78115 00:08:51.210 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 78115 00:08:51.210 [2024-12-07 05:36:24.552101] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:51.468 [2024-12-07 05:36:24.583957] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:51.468 ************************************ 00:08:51.468 END TEST raid_state_function_test 00:08:51.468 ************************************ 00:08:51.468 05:36:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:51.468 00:08:51.468 real 0m8.402s 00:08:51.468 user 0m14.476s 00:08:51.468 sys 0m1.629s 00:08:51.468 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:51.468 05:36:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.728 05:36:24 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:08:51.728 05:36:24 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:51.728 05:36:24 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:51.728 05:36:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:51.728 ************************************ 00:08:51.728 START TEST raid_state_function_test_sb 00:08:51.728 ************************************ 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 true 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:51.728 Process raid pid: 78714 00:08:51.728 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=78714 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78714' 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 78714 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 78714 ']' 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.728 05:36:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:51.728 [2024-12-07 05:36:24.951260] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:08:51.728 [2024-12-07 05:36:24.951518] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:51.987 [2024-12-07 05:36:25.106730] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:51.987 [2024-12-07 05:36:25.132380] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:51.987 [2024-12-07 05:36:25.174817] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:51.987 [2024-12-07 05:36:25.174931] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:52.554 05:36:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:52.554 05:36:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:08:52.554 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:52.554 05:36:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.555 [2024-12-07 05:36:25.809409] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:52.555 [2024-12-07 05:36:25.809521] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:52.555 [2024-12-07 05:36:25.809556] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:52.555 [2024-12-07 05:36:25.809601] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:52.555 [2024-12-07 05:36:25.809658] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:52.555 [2024-12-07 05:36:25.809697] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.555 "name": "Existed_Raid", 00:08:52.555 "uuid": "78d5f904-ef22-4181-8d3f-c9b73bf230a6", 00:08:52.555 "strip_size_kb": 0, 00:08:52.555 "state": "configuring", 00:08:52.555 "raid_level": "raid1", 00:08:52.555 "superblock": true, 00:08:52.555 "num_base_bdevs": 3, 00:08:52.555 "num_base_bdevs_discovered": 0, 00:08:52.555 "num_base_bdevs_operational": 3, 00:08:52.555 "base_bdevs_list": [ 00:08:52.555 { 00:08:52.555 "name": "BaseBdev1", 00:08:52.555 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.555 "is_configured": false, 00:08:52.555 "data_offset": 0, 00:08:52.555 "data_size": 0 00:08:52.555 }, 00:08:52.555 { 00:08:52.555 "name": "BaseBdev2", 00:08:52.555 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.555 "is_configured": false, 00:08:52.555 "data_offset": 0, 00:08:52.555 "data_size": 0 00:08:52.555 }, 00:08:52.555 { 00:08:52.555 "name": "BaseBdev3", 00:08:52.555 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.555 "is_configured": false, 00:08:52.555 "data_offset": 0, 00:08:52.555 "data_size": 0 00:08:52.555 } 00:08:52.555 ] 00:08:52.555 }' 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.555 05:36:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.123 [2024-12-07 05:36:26.268612] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:53.123 [2024-12-07 05:36:26.268675] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.123 [2024-12-07 05:36:26.280604] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:53.123 [2024-12-07 05:36:26.280677] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:53.123 [2024-12-07 05:36:26.280687] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:53.123 [2024-12-07 05:36:26.280696] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:53.123 [2024-12-07 05:36:26.280702] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:53.123 [2024-12-07 05:36:26.280711] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.123 [2024-12-07 05:36:26.301528] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:53.123 BaseBdev1 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.123 [ 00:08:53.123 { 00:08:53.123 "name": "BaseBdev1", 00:08:53.123 "aliases": [ 00:08:53.123 "57cc67b6-f63b-4500-94ee-df8890e90edb" 00:08:53.123 ], 00:08:53.123 "product_name": "Malloc disk", 00:08:53.123 "block_size": 512, 00:08:53.123 "num_blocks": 65536, 00:08:53.123 "uuid": "57cc67b6-f63b-4500-94ee-df8890e90edb", 00:08:53.123 "assigned_rate_limits": { 00:08:53.123 "rw_ios_per_sec": 0, 00:08:53.123 "rw_mbytes_per_sec": 0, 00:08:53.123 "r_mbytes_per_sec": 0, 00:08:53.123 "w_mbytes_per_sec": 0 00:08:53.123 }, 00:08:53.123 "claimed": true, 00:08:53.123 "claim_type": "exclusive_write", 00:08:53.123 "zoned": false, 00:08:53.123 "supported_io_types": { 00:08:53.123 "read": true, 00:08:53.123 "write": true, 00:08:53.123 "unmap": true, 00:08:53.123 "flush": true, 00:08:53.123 "reset": true, 00:08:53.123 "nvme_admin": false, 00:08:53.123 "nvme_io": false, 00:08:53.123 "nvme_io_md": false, 00:08:53.123 "write_zeroes": true, 00:08:53.123 "zcopy": true, 00:08:53.123 "get_zone_info": false, 00:08:53.123 "zone_management": false, 00:08:53.123 "zone_append": false, 00:08:53.123 "compare": false, 00:08:53.123 "compare_and_write": false, 00:08:53.123 "abort": true, 00:08:53.123 "seek_hole": false, 00:08:53.123 "seek_data": false, 00:08:53.123 "copy": true, 00:08:53.123 "nvme_iov_md": false 00:08:53.123 }, 00:08:53.123 "memory_domains": [ 00:08:53.123 { 00:08:53.123 "dma_device_id": "system", 00:08:53.123 "dma_device_type": 1 00:08:53.123 }, 00:08:53.123 { 00:08:53.123 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:53.123 "dma_device_type": 2 00:08:53.123 } 00:08:53.123 ], 00:08:53.123 "driver_specific": {} 00:08:53.123 } 00:08:53.123 ] 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.123 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:53.124 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.124 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.124 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.124 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.124 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.124 "name": "Existed_Raid", 00:08:53.124 "uuid": "d81956e5-32f4-4d86-9b31-f8b5248f829d", 00:08:53.124 "strip_size_kb": 0, 00:08:53.124 "state": "configuring", 00:08:53.124 "raid_level": "raid1", 00:08:53.124 "superblock": true, 00:08:53.124 "num_base_bdevs": 3, 00:08:53.124 "num_base_bdevs_discovered": 1, 00:08:53.124 "num_base_bdevs_operational": 3, 00:08:53.124 "base_bdevs_list": [ 00:08:53.124 { 00:08:53.124 "name": "BaseBdev1", 00:08:53.124 "uuid": "57cc67b6-f63b-4500-94ee-df8890e90edb", 00:08:53.124 "is_configured": true, 00:08:53.124 "data_offset": 2048, 00:08:53.124 "data_size": 63488 00:08:53.124 }, 00:08:53.124 { 00:08:53.124 "name": "BaseBdev2", 00:08:53.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:53.124 "is_configured": false, 00:08:53.124 "data_offset": 0, 00:08:53.124 "data_size": 0 00:08:53.124 }, 00:08:53.124 { 00:08:53.124 "name": "BaseBdev3", 00:08:53.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:53.124 "is_configured": false, 00:08:53.124 "data_offset": 0, 00:08:53.124 "data_size": 0 00:08:53.124 } 00:08:53.124 ] 00:08:53.124 }' 00:08:53.124 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.124 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.692 [2024-12-07 05:36:26.772800] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:53.692 [2024-12-07 05:36:26.772927] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.692 [2024-12-07 05:36:26.784826] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:53.692 [2024-12-07 05:36:26.786735] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:53.692 [2024-12-07 05:36:26.786774] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:53.692 [2024-12-07 05:36:26.786784] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:53.692 [2024-12-07 05:36:26.786794] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.692 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.692 "name": "Existed_Raid", 00:08:53.692 "uuid": "cb84530f-e458-4bc4-8fe2-afc8b7b6ea23", 00:08:53.692 "strip_size_kb": 0, 00:08:53.692 "state": "configuring", 00:08:53.692 "raid_level": "raid1", 00:08:53.692 "superblock": true, 00:08:53.692 "num_base_bdevs": 3, 00:08:53.692 "num_base_bdevs_discovered": 1, 00:08:53.692 "num_base_bdevs_operational": 3, 00:08:53.692 "base_bdevs_list": [ 00:08:53.692 { 00:08:53.692 "name": "BaseBdev1", 00:08:53.692 "uuid": "57cc67b6-f63b-4500-94ee-df8890e90edb", 00:08:53.692 "is_configured": true, 00:08:53.693 "data_offset": 2048, 00:08:53.693 "data_size": 63488 00:08:53.693 }, 00:08:53.693 { 00:08:53.693 "name": "BaseBdev2", 00:08:53.693 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:53.693 "is_configured": false, 00:08:53.693 "data_offset": 0, 00:08:53.693 "data_size": 0 00:08:53.693 }, 00:08:53.693 { 00:08:53.693 "name": "BaseBdev3", 00:08:53.693 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:53.693 "is_configured": false, 00:08:53.693 "data_offset": 0, 00:08:53.693 "data_size": 0 00:08:53.693 } 00:08:53.693 ] 00:08:53.693 }' 00:08:53.693 05:36:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.693 05:36:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.952 [2024-12-07 05:36:27.227069] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:53.952 BaseBdev2 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.952 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.952 [ 00:08:53.952 { 00:08:53.952 "name": "BaseBdev2", 00:08:53.953 "aliases": [ 00:08:53.953 "d5155b7a-2500-4fc5-b9bb-ab2e6c5328fd" 00:08:53.953 ], 00:08:53.953 "product_name": "Malloc disk", 00:08:53.953 "block_size": 512, 00:08:53.953 "num_blocks": 65536, 00:08:53.953 "uuid": "d5155b7a-2500-4fc5-b9bb-ab2e6c5328fd", 00:08:53.953 "assigned_rate_limits": { 00:08:53.953 "rw_ios_per_sec": 0, 00:08:53.953 "rw_mbytes_per_sec": 0, 00:08:53.953 "r_mbytes_per_sec": 0, 00:08:53.953 "w_mbytes_per_sec": 0 00:08:53.953 }, 00:08:53.953 "claimed": true, 00:08:53.953 "claim_type": "exclusive_write", 00:08:53.953 "zoned": false, 00:08:53.953 "supported_io_types": { 00:08:53.953 "read": true, 00:08:53.953 "write": true, 00:08:53.953 "unmap": true, 00:08:53.953 "flush": true, 00:08:53.953 "reset": true, 00:08:53.953 "nvme_admin": false, 00:08:53.953 "nvme_io": false, 00:08:53.953 "nvme_io_md": false, 00:08:53.953 "write_zeroes": true, 00:08:53.953 "zcopy": true, 00:08:53.953 "get_zone_info": false, 00:08:53.953 "zone_management": false, 00:08:53.953 "zone_append": false, 00:08:53.953 "compare": false, 00:08:53.953 "compare_and_write": false, 00:08:53.953 "abort": true, 00:08:53.953 "seek_hole": false, 00:08:53.953 "seek_data": false, 00:08:53.953 "copy": true, 00:08:53.953 "nvme_iov_md": false 00:08:53.953 }, 00:08:53.953 "memory_domains": [ 00:08:53.953 { 00:08:53.953 "dma_device_id": "system", 00:08:53.953 "dma_device_type": 1 00:08:53.953 }, 00:08:53.953 { 00:08:53.953 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:53.953 "dma_device_type": 2 00:08:53.953 } 00:08:53.953 ], 00:08:53.953 "driver_specific": {} 00:08:53.953 } 00:08:53.953 ] 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.953 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.213 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.213 "name": "Existed_Raid", 00:08:54.213 "uuid": "cb84530f-e458-4bc4-8fe2-afc8b7b6ea23", 00:08:54.213 "strip_size_kb": 0, 00:08:54.213 "state": "configuring", 00:08:54.213 "raid_level": "raid1", 00:08:54.213 "superblock": true, 00:08:54.213 "num_base_bdevs": 3, 00:08:54.213 "num_base_bdevs_discovered": 2, 00:08:54.213 "num_base_bdevs_operational": 3, 00:08:54.213 "base_bdevs_list": [ 00:08:54.213 { 00:08:54.213 "name": "BaseBdev1", 00:08:54.213 "uuid": "57cc67b6-f63b-4500-94ee-df8890e90edb", 00:08:54.213 "is_configured": true, 00:08:54.213 "data_offset": 2048, 00:08:54.213 "data_size": 63488 00:08:54.213 }, 00:08:54.213 { 00:08:54.213 "name": "BaseBdev2", 00:08:54.213 "uuid": "d5155b7a-2500-4fc5-b9bb-ab2e6c5328fd", 00:08:54.213 "is_configured": true, 00:08:54.213 "data_offset": 2048, 00:08:54.213 "data_size": 63488 00:08:54.213 }, 00:08:54.213 { 00:08:54.213 "name": "BaseBdev3", 00:08:54.213 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:54.213 "is_configured": false, 00:08:54.213 "data_offset": 0, 00:08:54.213 "data_size": 0 00:08:54.213 } 00:08:54.213 ] 00:08:54.213 }' 00:08:54.213 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.213 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.473 [2024-12-07 05:36:27.781661] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:54.473 [2024-12-07 05:36:27.781984] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:54.473 [2024-12-07 05:36:27.782012] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:54.473 [2024-12-07 05:36:27.782359] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:54.473 BaseBdev3 00:08:54.473 [2024-12-07 05:36:27.782558] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:54.473 [2024-12-07 05:36:27.782576] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:54.473 [2024-12-07 05:36:27.782748] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.473 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.473 [ 00:08:54.473 { 00:08:54.474 "name": "BaseBdev3", 00:08:54.474 "aliases": [ 00:08:54.474 "3f08d5f4-0162-4398-b053-e203e2909bdf" 00:08:54.474 ], 00:08:54.474 "product_name": "Malloc disk", 00:08:54.474 "block_size": 512, 00:08:54.474 "num_blocks": 65536, 00:08:54.474 "uuid": "3f08d5f4-0162-4398-b053-e203e2909bdf", 00:08:54.474 "assigned_rate_limits": { 00:08:54.474 "rw_ios_per_sec": 0, 00:08:54.474 "rw_mbytes_per_sec": 0, 00:08:54.474 "r_mbytes_per_sec": 0, 00:08:54.474 "w_mbytes_per_sec": 0 00:08:54.474 }, 00:08:54.474 "claimed": true, 00:08:54.474 "claim_type": "exclusive_write", 00:08:54.474 "zoned": false, 00:08:54.474 "supported_io_types": { 00:08:54.474 "read": true, 00:08:54.474 "write": true, 00:08:54.474 "unmap": true, 00:08:54.474 "flush": true, 00:08:54.474 "reset": true, 00:08:54.474 "nvme_admin": false, 00:08:54.474 "nvme_io": false, 00:08:54.474 "nvme_io_md": false, 00:08:54.474 "write_zeroes": true, 00:08:54.474 "zcopy": true, 00:08:54.474 "get_zone_info": false, 00:08:54.474 "zone_management": false, 00:08:54.474 "zone_append": false, 00:08:54.474 "compare": false, 00:08:54.474 "compare_and_write": false, 00:08:54.474 "abort": true, 00:08:54.474 "seek_hole": false, 00:08:54.474 "seek_data": false, 00:08:54.474 "copy": true, 00:08:54.474 "nvme_iov_md": false 00:08:54.474 }, 00:08:54.474 "memory_domains": [ 00:08:54.474 { 00:08:54.474 "dma_device_id": "system", 00:08:54.474 "dma_device_type": 1 00:08:54.474 }, 00:08:54.474 { 00:08:54.474 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:54.474 "dma_device_type": 2 00:08:54.474 } 00:08:54.474 ], 00:08:54.474 "driver_specific": {} 00:08:54.474 } 00:08:54.474 ] 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.474 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.732 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.732 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.732 "name": "Existed_Raid", 00:08:54.732 "uuid": "cb84530f-e458-4bc4-8fe2-afc8b7b6ea23", 00:08:54.732 "strip_size_kb": 0, 00:08:54.732 "state": "online", 00:08:54.732 "raid_level": "raid1", 00:08:54.732 "superblock": true, 00:08:54.732 "num_base_bdevs": 3, 00:08:54.732 "num_base_bdevs_discovered": 3, 00:08:54.732 "num_base_bdevs_operational": 3, 00:08:54.732 "base_bdevs_list": [ 00:08:54.732 { 00:08:54.732 "name": "BaseBdev1", 00:08:54.732 "uuid": "57cc67b6-f63b-4500-94ee-df8890e90edb", 00:08:54.732 "is_configured": true, 00:08:54.732 "data_offset": 2048, 00:08:54.732 "data_size": 63488 00:08:54.732 }, 00:08:54.732 { 00:08:54.732 "name": "BaseBdev2", 00:08:54.732 "uuid": "d5155b7a-2500-4fc5-b9bb-ab2e6c5328fd", 00:08:54.732 "is_configured": true, 00:08:54.732 "data_offset": 2048, 00:08:54.732 "data_size": 63488 00:08:54.732 }, 00:08:54.732 { 00:08:54.732 "name": "BaseBdev3", 00:08:54.732 "uuid": "3f08d5f4-0162-4398-b053-e203e2909bdf", 00:08:54.732 "is_configured": true, 00:08:54.733 "data_offset": 2048, 00:08:54.733 "data_size": 63488 00:08:54.733 } 00:08:54.733 ] 00:08:54.733 }' 00:08:54.733 05:36:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.733 05:36:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.991 [2024-12-07 05:36:28.289120] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.991 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:54.991 "name": "Existed_Raid", 00:08:54.991 "aliases": [ 00:08:54.991 "cb84530f-e458-4bc4-8fe2-afc8b7b6ea23" 00:08:54.991 ], 00:08:54.991 "product_name": "Raid Volume", 00:08:54.991 "block_size": 512, 00:08:54.991 "num_blocks": 63488, 00:08:54.991 "uuid": "cb84530f-e458-4bc4-8fe2-afc8b7b6ea23", 00:08:54.991 "assigned_rate_limits": { 00:08:54.991 "rw_ios_per_sec": 0, 00:08:54.991 "rw_mbytes_per_sec": 0, 00:08:54.991 "r_mbytes_per_sec": 0, 00:08:54.991 "w_mbytes_per_sec": 0 00:08:54.991 }, 00:08:54.991 "claimed": false, 00:08:54.991 "zoned": false, 00:08:54.991 "supported_io_types": { 00:08:54.991 "read": true, 00:08:54.991 "write": true, 00:08:54.991 "unmap": false, 00:08:54.991 "flush": false, 00:08:54.991 "reset": true, 00:08:54.991 "nvme_admin": false, 00:08:54.991 "nvme_io": false, 00:08:54.991 "nvme_io_md": false, 00:08:54.991 "write_zeroes": true, 00:08:54.991 "zcopy": false, 00:08:54.991 "get_zone_info": false, 00:08:54.991 "zone_management": false, 00:08:54.991 "zone_append": false, 00:08:54.991 "compare": false, 00:08:54.991 "compare_and_write": false, 00:08:54.991 "abort": false, 00:08:54.991 "seek_hole": false, 00:08:54.991 "seek_data": false, 00:08:54.991 "copy": false, 00:08:54.991 "nvme_iov_md": false 00:08:54.991 }, 00:08:54.991 "memory_domains": [ 00:08:54.991 { 00:08:54.991 "dma_device_id": "system", 00:08:54.991 "dma_device_type": 1 00:08:54.991 }, 00:08:54.991 { 00:08:54.991 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:54.991 "dma_device_type": 2 00:08:54.991 }, 00:08:54.991 { 00:08:54.991 "dma_device_id": "system", 00:08:54.991 "dma_device_type": 1 00:08:54.991 }, 00:08:54.991 { 00:08:54.991 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:54.991 "dma_device_type": 2 00:08:54.991 }, 00:08:54.991 { 00:08:54.991 "dma_device_id": "system", 00:08:54.991 "dma_device_type": 1 00:08:54.991 }, 00:08:54.991 { 00:08:54.991 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:54.991 "dma_device_type": 2 00:08:54.991 } 00:08:54.991 ], 00:08:54.991 "driver_specific": { 00:08:54.991 "raid": { 00:08:54.991 "uuid": "cb84530f-e458-4bc4-8fe2-afc8b7b6ea23", 00:08:54.991 "strip_size_kb": 0, 00:08:54.991 "state": "online", 00:08:54.991 "raid_level": "raid1", 00:08:54.991 "superblock": true, 00:08:54.991 "num_base_bdevs": 3, 00:08:54.991 "num_base_bdevs_discovered": 3, 00:08:54.991 "num_base_bdevs_operational": 3, 00:08:54.991 "base_bdevs_list": [ 00:08:54.991 { 00:08:54.991 "name": "BaseBdev1", 00:08:54.991 "uuid": "57cc67b6-f63b-4500-94ee-df8890e90edb", 00:08:54.991 "is_configured": true, 00:08:54.991 "data_offset": 2048, 00:08:54.991 "data_size": 63488 00:08:54.991 }, 00:08:54.991 { 00:08:54.991 "name": "BaseBdev2", 00:08:54.991 "uuid": "d5155b7a-2500-4fc5-b9bb-ab2e6c5328fd", 00:08:54.991 "is_configured": true, 00:08:54.991 "data_offset": 2048, 00:08:54.991 "data_size": 63488 00:08:54.992 }, 00:08:54.992 { 00:08:54.992 "name": "BaseBdev3", 00:08:54.992 "uuid": "3f08d5f4-0162-4398-b053-e203e2909bdf", 00:08:54.992 "is_configured": true, 00:08:54.992 "data_offset": 2048, 00:08:54.992 "data_size": 63488 00:08:54.992 } 00:08:54.992 ] 00:08:54.992 } 00:08:54.992 } 00:08:54.992 }' 00:08:54.992 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:54.992 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:54.992 BaseBdev2 00:08:54.992 BaseBdev3' 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.250 [2024-12-07 05:36:28.564405] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:55.250 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:55.251 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:55.251 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:55.251 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:55.251 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:55.251 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:55.251 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:55.251 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:55.251 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.251 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.251 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.251 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.510 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:55.510 "name": "Existed_Raid", 00:08:55.510 "uuid": "cb84530f-e458-4bc4-8fe2-afc8b7b6ea23", 00:08:55.510 "strip_size_kb": 0, 00:08:55.510 "state": "online", 00:08:55.510 "raid_level": "raid1", 00:08:55.510 "superblock": true, 00:08:55.510 "num_base_bdevs": 3, 00:08:55.510 "num_base_bdevs_discovered": 2, 00:08:55.510 "num_base_bdevs_operational": 2, 00:08:55.510 "base_bdevs_list": [ 00:08:55.510 { 00:08:55.510 "name": null, 00:08:55.510 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:55.510 "is_configured": false, 00:08:55.510 "data_offset": 0, 00:08:55.510 "data_size": 63488 00:08:55.510 }, 00:08:55.510 { 00:08:55.510 "name": "BaseBdev2", 00:08:55.510 "uuid": "d5155b7a-2500-4fc5-b9bb-ab2e6c5328fd", 00:08:55.510 "is_configured": true, 00:08:55.510 "data_offset": 2048, 00:08:55.510 "data_size": 63488 00:08:55.510 }, 00:08:55.510 { 00:08:55.510 "name": "BaseBdev3", 00:08:55.510 "uuid": "3f08d5f4-0162-4398-b053-e203e2909bdf", 00:08:55.510 "is_configured": true, 00:08:55.510 "data_offset": 2048, 00:08:55.510 "data_size": 63488 00:08:55.510 } 00:08:55.510 ] 00:08:55.510 }' 00:08:55.510 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:55.510 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.769 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:55.769 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:55.769 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.769 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.769 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.769 05:36:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:55.770 05:36:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.770 [2024-12-07 05:36:29.031173] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.770 [2024-12-07 05:36:29.102753] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:55.770 [2024-12-07 05:36:29.102921] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:55.770 [2024-12-07 05:36:29.114473] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:55.770 [2024-12-07 05:36:29.114593] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:55.770 [2024-12-07 05:36:29.114634] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.770 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.029 BaseBdev2 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.029 [ 00:08:56.029 { 00:08:56.029 "name": "BaseBdev2", 00:08:56.029 "aliases": [ 00:08:56.029 "1cc627b0-7767-4feb-9760-b3dbc3be7cfd" 00:08:56.029 ], 00:08:56.029 "product_name": "Malloc disk", 00:08:56.029 "block_size": 512, 00:08:56.029 "num_blocks": 65536, 00:08:56.029 "uuid": "1cc627b0-7767-4feb-9760-b3dbc3be7cfd", 00:08:56.029 "assigned_rate_limits": { 00:08:56.029 "rw_ios_per_sec": 0, 00:08:56.029 "rw_mbytes_per_sec": 0, 00:08:56.029 "r_mbytes_per_sec": 0, 00:08:56.029 "w_mbytes_per_sec": 0 00:08:56.029 }, 00:08:56.029 "claimed": false, 00:08:56.029 "zoned": false, 00:08:56.029 "supported_io_types": { 00:08:56.029 "read": true, 00:08:56.029 "write": true, 00:08:56.029 "unmap": true, 00:08:56.029 "flush": true, 00:08:56.029 "reset": true, 00:08:56.029 "nvme_admin": false, 00:08:56.029 "nvme_io": false, 00:08:56.029 "nvme_io_md": false, 00:08:56.029 "write_zeroes": true, 00:08:56.029 "zcopy": true, 00:08:56.029 "get_zone_info": false, 00:08:56.029 "zone_management": false, 00:08:56.029 "zone_append": false, 00:08:56.029 "compare": false, 00:08:56.029 "compare_and_write": false, 00:08:56.029 "abort": true, 00:08:56.029 "seek_hole": false, 00:08:56.029 "seek_data": false, 00:08:56.029 "copy": true, 00:08:56.029 "nvme_iov_md": false 00:08:56.029 }, 00:08:56.029 "memory_domains": [ 00:08:56.029 { 00:08:56.029 "dma_device_id": "system", 00:08:56.029 "dma_device_type": 1 00:08:56.029 }, 00:08:56.029 { 00:08:56.029 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:56.029 "dma_device_type": 2 00:08:56.029 } 00:08:56.029 ], 00:08:56.029 "driver_specific": {} 00:08:56.029 } 00:08:56.029 ] 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.029 BaseBdev3 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.029 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.029 [ 00:08:56.029 { 00:08:56.029 "name": "BaseBdev3", 00:08:56.029 "aliases": [ 00:08:56.029 "f72a6bcb-63f4-4b4f-8d60-73d0a1b98de7" 00:08:56.029 ], 00:08:56.029 "product_name": "Malloc disk", 00:08:56.029 "block_size": 512, 00:08:56.029 "num_blocks": 65536, 00:08:56.029 "uuid": "f72a6bcb-63f4-4b4f-8d60-73d0a1b98de7", 00:08:56.029 "assigned_rate_limits": { 00:08:56.029 "rw_ios_per_sec": 0, 00:08:56.029 "rw_mbytes_per_sec": 0, 00:08:56.029 "r_mbytes_per_sec": 0, 00:08:56.029 "w_mbytes_per_sec": 0 00:08:56.029 }, 00:08:56.029 "claimed": false, 00:08:56.029 "zoned": false, 00:08:56.029 "supported_io_types": { 00:08:56.029 "read": true, 00:08:56.029 "write": true, 00:08:56.029 "unmap": true, 00:08:56.030 "flush": true, 00:08:56.030 "reset": true, 00:08:56.030 "nvme_admin": false, 00:08:56.030 "nvme_io": false, 00:08:56.030 "nvme_io_md": false, 00:08:56.030 "write_zeroes": true, 00:08:56.030 "zcopy": true, 00:08:56.030 "get_zone_info": false, 00:08:56.030 "zone_management": false, 00:08:56.030 "zone_append": false, 00:08:56.030 "compare": false, 00:08:56.030 "compare_and_write": false, 00:08:56.030 "abort": true, 00:08:56.030 "seek_hole": false, 00:08:56.030 "seek_data": false, 00:08:56.030 "copy": true, 00:08:56.030 "nvme_iov_md": false 00:08:56.030 }, 00:08:56.030 "memory_domains": [ 00:08:56.030 { 00:08:56.030 "dma_device_id": "system", 00:08:56.030 "dma_device_type": 1 00:08:56.030 }, 00:08:56.030 { 00:08:56.030 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:56.030 "dma_device_type": 2 00:08:56.030 } 00:08:56.030 ], 00:08:56.030 "driver_specific": {} 00:08:56.030 } 00:08:56.030 ] 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.030 [2024-12-07 05:36:29.278223] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:56.030 [2024-12-07 05:36:29.278308] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:56.030 [2024-12-07 05:36:29.278360] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:56.030 [2024-12-07 05:36:29.280242] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.030 "name": "Existed_Raid", 00:08:56.030 "uuid": "6fc5de2c-a9ea-4ce2-b773-51d0970a1cb5", 00:08:56.030 "strip_size_kb": 0, 00:08:56.030 "state": "configuring", 00:08:56.030 "raid_level": "raid1", 00:08:56.030 "superblock": true, 00:08:56.030 "num_base_bdevs": 3, 00:08:56.030 "num_base_bdevs_discovered": 2, 00:08:56.030 "num_base_bdevs_operational": 3, 00:08:56.030 "base_bdevs_list": [ 00:08:56.030 { 00:08:56.030 "name": "BaseBdev1", 00:08:56.030 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.030 "is_configured": false, 00:08:56.030 "data_offset": 0, 00:08:56.030 "data_size": 0 00:08:56.030 }, 00:08:56.030 { 00:08:56.030 "name": "BaseBdev2", 00:08:56.030 "uuid": "1cc627b0-7767-4feb-9760-b3dbc3be7cfd", 00:08:56.030 "is_configured": true, 00:08:56.030 "data_offset": 2048, 00:08:56.030 "data_size": 63488 00:08:56.030 }, 00:08:56.030 { 00:08:56.030 "name": "BaseBdev3", 00:08:56.030 "uuid": "f72a6bcb-63f4-4b4f-8d60-73d0a1b98de7", 00:08:56.030 "is_configured": true, 00:08:56.030 "data_offset": 2048, 00:08:56.030 "data_size": 63488 00:08:56.030 } 00:08:56.030 ] 00:08:56.030 }' 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.030 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.598 [2024-12-07 05:36:29.749457] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.598 "name": "Existed_Raid", 00:08:56.598 "uuid": "6fc5de2c-a9ea-4ce2-b773-51d0970a1cb5", 00:08:56.598 "strip_size_kb": 0, 00:08:56.598 "state": "configuring", 00:08:56.598 "raid_level": "raid1", 00:08:56.598 "superblock": true, 00:08:56.598 "num_base_bdevs": 3, 00:08:56.598 "num_base_bdevs_discovered": 1, 00:08:56.598 "num_base_bdevs_operational": 3, 00:08:56.598 "base_bdevs_list": [ 00:08:56.598 { 00:08:56.598 "name": "BaseBdev1", 00:08:56.598 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.598 "is_configured": false, 00:08:56.598 "data_offset": 0, 00:08:56.598 "data_size": 0 00:08:56.598 }, 00:08:56.598 { 00:08:56.598 "name": null, 00:08:56.598 "uuid": "1cc627b0-7767-4feb-9760-b3dbc3be7cfd", 00:08:56.598 "is_configured": false, 00:08:56.598 "data_offset": 0, 00:08:56.598 "data_size": 63488 00:08:56.598 }, 00:08:56.598 { 00:08:56.598 "name": "BaseBdev3", 00:08:56.598 "uuid": "f72a6bcb-63f4-4b4f-8d60-73d0a1b98de7", 00:08:56.598 "is_configured": true, 00:08:56.598 "data_offset": 2048, 00:08:56.598 "data_size": 63488 00:08:56.598 } 00:08:56.598 ] 00:08:56.598 }' 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.598 05:36:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.857 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.857 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.857 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.857 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:56.857 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.857 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:56.857 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:56.857 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.857 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.115 [2024-12-07 05:36:30.231512] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:57.115 BaseBdev1 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.115 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.115 [ 00:08:57.115 { 00:08:57.115 "name": "BaseBdev1", 00:08:57.115 "aliases": [ 00:08:57.115 "3783ade2-5aec-44e6-b0c0-7ffadf644661" 00:08:57.115 ], 00:08:57.115 "product_name": "Malloc disk", 00:08:57.115 "block_size": 512, 00:08:57.115 "num_blocks": 65536, 00:08:57.115 "uuid": "3783ade2-5aec-44e6-b0c0-7ffadf644661", 00:08:57.115 "assigned_rate_limits": { 00:08:57.115 "rw_ios_per_sec": 0, 00:08:57.115 "rw_mbytes_per_sec": 0, 00:08:57.115 "r_mbytes_per_sec": 0, 00:08:57.115 "w_mbytes_per_sec": 0 00:08:57.115 }, 00:08:57.115 "claimed": true, 00:08:57.115 "claim_type": "exclusive_write", 00:08:57.115 "zoned": false, 00:08:57.115 "supported_io_types": { 00:08:57.115 "read": true, 00:08:57.115 "write": true, 00:08:57.115 "unmap": true, 00:08:57.115 "flush": true, 00:08:57.115 "reset": true, 00:08:57.115 "nvme_admin": false, 00:08:57.115 "nvme_io": false, 00:08:57.115 "nvme_io_md": false, 00:08:57.116 "write_zeroes": true, 00:08:57.116 "zcopy": true, 00:08:57.116 "get_zone_info": false, 00:08:57.116 "zone_management": false, 00:08:57.116 "zone_append": false, 00:08:57.116 "compare": false, 00:08:57.116 "compare_and_write": false, 00:08:57.116 "abort": true, 00:08:57.116 "seek_hole": false, 00:08:57.116 "seek_data": false, 00:08:57.116 "copy": true, 00:08:57.116 "nvme_iov_md": false 00:08:57.116 }, 00:08:57.116 "memory_domains": [ 00:08:57.116 { 00:08:57.116 "dma_device_id": "system", 00:08:57.116 "dma_device_type": 1 00:08:57.116 }, 00:08:57.116 { 00:08:57.116 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:57.116 "dma_device_type": 2 00:08:57.116 } 00:08:57.116 ], 00:08:57.116 "driver_specific": {} 00:08:57.116 } 00:08:57.116 ] 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.116 "name": "Existed_Raid", 00:08:57.116 "uuid": "6fc5de2c-a9ea-4ce2-b773-51d0970a1cb5", 00:08:57.116 "strip_size_kb": 0, 00:08:57.116 "state": "configuring", 00:08:57.116 "raid_level": "raid1", 00:08:57.116 "superblock": true, 00:08:57.116 "num_base_bdevs": 3, 00:08:57.116 "num_base_bdevs_discovered": 2, 00:08:57.116 "num_base_bdevs_operational": 3, 00:08:57.116 "base_bdevs_list": [ 00:08:57.116 { 00:08:57.116 "name": "BaseBdev1", 00:08:57.116 "uuid": "3783ade2-5aec-44e6-b0c0-7ffadf644661", 00:08:57.116 "is_configured": true, 00:08:57.116 "data_offset": 2048, 00:08:57.116 "data_size": 63488 00:08:57.116 }, 00:08:57.116 { 00:08:57.116 "name": null, 00:08:57.116 "uuid": "1cc627b0-7767-4feb-9760-b3dbc3be7cfd", 00:08:57.116 "is_configured": false, 00:08:57.116 "data_offset": 0, 00:08:57.116 "data_size": 63488 00:08:57.116 }, 00:08:57.116 { 00:08:57.116 "name": "BaseBdev3", 00:08:57.116 "uuid": "f72a6bcb-63f4-4b4f-8d60-73d0a1b98de7", 00:08:57.116 "is_configured": true, 00:08:57.116 "data_offset": 2048, 00:08:57.116 "data_size": 63488 00:08:57.116 } 00:08:57.116 ] 00:08:57.116 }' 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.116 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.374 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.374 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:57.374 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.374 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.374 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.374 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:57.374 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:57.374 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.374 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.631 [2024-12-07 05:36:30.742752] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.631 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.632 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.632 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.632 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.632 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.632 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.632 "name": "Existed_Raid", 00:08:57.632 "uuid": "6fc5de2c-a9ea-4ce2-b773-51d0970a1cb5", 00:08:57.632 "strip_size_kb": 0, 00:08:57.632 "state": "configuring", 00:08:57.632 "raid_level": "raid1", 00:08:57.632 "superblock": true, 00:08:57.632 "num_base_bdevs": 3, 00:08:57.632 "num_base_bdevs_discovered": 1, 00:08:57.632 "num_base_bdevs_operational": 3, 00:08:57.632 "base_bdevs_list": [ 00:08:57.632 { 00:08:57.632 "name": "BaseBdev1", 00:08:57.632 "uuid": "3783ade2-5aec-44e6-b0c0-7ffadf644661", 00:08:57.632 "is_configured": true, 00:08:57.632 "data_offset": 2048, 00:08:57.632 "data_size": 63488 00:08:57.632 }, 00:08:57.632 { 00:08:57.632 "name": null, 00:08:57.632 "uuid": "1cc627b0-7767-4feb-9760-b3dbc3be7cfd", 00:08:57.632 "is_configured": false, 00:08:57.632 "data_offset": 0, 00:08:57.632 "data_size": 63488 00:08:57.632 }, 00:08:57.632 { 00:08:57.632 "name": null, 00:08:57.632 "uuid": "f72a6bcb-63f4-4b4f-8d60-73d0a1b98de7", 00:08:57.632 "is_configured": false, 00:08:57.632 "data_offset": 0, 00:08:57.632 "data_size": 63488 00:08:57.632 } 00:08:57.632 ] 00:08:57.632 }' 00:08:57.632 05:36:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.632 05:36:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:57.889 [2024-12-07 05:36:31.241905] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.889 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:57.890 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:57.890 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:57.890 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:57.890 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.890 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.890 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.890 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.890 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.890 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.890 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.890 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.148 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.148 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.148 "name": "Existed_Raid", 00:08:58.148 "uuid": "6fc5de2c-a9ea-4ce2-b773-51d0970a1cb5", 00:08:58.148 "strip_size_kb": 0, 00:08:58.148 "state": "configuring", 00:08:58.148 "raid_level": "raid1", 00:08:58.148 "superblock": true, 00:08:58.148 "num_base_bdevs": 3, 00:08:58.148 "num_base_bdevs_discovered": 2, 00:08:58.148 "num_base_bdevs_operational": 3, 00:08:58.148 "base_bdevs_list": [ 00:08:58.148 { 00:08:58.148 "name": "BaseBdev1", 00:08:58.148 "uuid": "3783ade2-5aec-44e6-b0c0-7ffadf644661", 00:08:58.148 "is_configured": true, 00:08:58.148 "data_offset": 2048, 00:08:58.148 "data_size": 63488 00:08:58.148 }, 00:08:58.148 { 00:08:58.148 "name": null, 00:08:58.148 "uuid": "1cc627b0-7767-4feb-9760-b3dbc3be7cfd", 00:08:58.148 "is_configured": false, 00:08:58.148 "data_offset": 0, 00:08:58.148 "data_size": 63488 00:08:58.148 }, 00:08:58.148 { 00:08:58.148 "name": "BaseBdev3", 00:08:58.148 "uuid": "f72a6bcb-63f4-4b4f-8d60-73d0a1b98de7", 00:08:58.148 "is_configured": true, 00:08:58.148 "data_offset": 2048, 00:08:58.148 "data_size": 63488 00:08:58.148 } 00:08:58.148 ] 00:08:58.148 }' 00:08:58.148 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.148 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.406 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.406 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:58.406 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.406 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.406 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.406 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.407 [2024-12-07 05:36:31.713124] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.407 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.666 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.666 "name": "Existed_Raid", 00:08:58.666 "uuid": "6fc5de2c-a9ea-4ce2-b773-51d0970a1cb5", 00:08:58.666 "strip_size_kb": 0, 00:08:58.666 "state": "configuring", 00:08:58.666 "raid_level": "raid1", 00:08:58.666 "superblock": true, 00:08:58.666 "num_base_bdevs": 3, 00:08:58.666 "num_base_bdevs_discovered": 1, 00:08:58.666 "num_base_bdevs_operational": 3, 00:08:58.666 "base_bdevs_list": [ 00:08:58.666 { 00:08:58.666 "name": null, 00:08:58.666 "uuid": "3783ade2-5aec-44e6-b0c0-7ffadf644661", 00:08:58.666 "is_configured": false, 00:08:58.666 "data_offset": 0, 00:08:58.666 "data_size": 63488 00:08:58.666 }, 00:08:58.666 { 00:08:58.666 "name": null, 00:08:58.666 "uuid": "1cc627b0-7767-4feb-9760-b3dbc3be7cfd", 00:08:58.666 "is_configured": false, 00:08:58.666 "data_offset": 0, 00:08:58.666 "data_size": 63488 00:08:58.666 }, 00:08:58.666 { 00:08:58.666 "name": "BaseBdev3", 00:08:58.666 "uuid": "f72a6bcb-63f4-4b4f-8d60-73d0a1b98de7", 00:08:58.666 "is_configured": true, 00:08:58.666 "data_offset": 2048, 00:08:58.666 "data_size": 63488 00:08:58.666 } 00:08:58.666 ] 00:08:58.666 }' 00:08:58.666 05:36:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.666 05:36:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.924 [2024-12-07 05:36:32.218788] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.924 "name": "Existed_Raid", 00:08:58.924 "uuid": "6fc5de2c-a9ea-4ce2-b773-51d0970a1cb5", 00:08:58.924 "strip_size_kb": 0, 00:08:58.924 "state": "configuring", 00:08:58.924 "raid_level": "raid1", 00:08:58.924 "superblock": true, 00:08:58.924 "num_base_bdevs": 3, 00:08:58.924 "num_base_bdevs_discovered": 2, 00:08:58.924 "num_base_bdevs_operational": 3, 00:08:58.924 "base_bdevs_list": [ 00:08:58.924 { 00:08:58.924 "name": null, 00:08:58.924 "uuid": "3783ade2-5aec-44e6-b0c0-7ffadf644661", 00:08:58.924 "is_configured": false, 00:08:58.924 "data_offset": 0, 00:08:58.924 "data_size": 63488 00:08:58.924 }, 00:08:58.924 { 00:08:58.924 "name": "BaseBdev2", 00:08:58.924 "uuid": "1cc627b0-7767-4feb-9760-b3dbc3be7cfd", 00:08:58.924 "is_configured": true, 00:08:58.924 "data_offset": 2048, 00:08:58.924 "data_size": 63488 00:08:58.924 }, 00:08:58.924 { 00:08:58.924 "name": "BaseBdev3", 00:08:58.924 "uuid": "f72a6bcb-63f4-4b4f-8d60-73d0a1b98de7", 00:08:58.924 "is_configured": true, 00:08:58.924 "data_offset": 2048, 00:08:58.924 "data_size": 63488 00:08:58.924 } 00:08:58.924 ] 00:08:58.924 }' 00:08:58.924 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.925 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.491 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 3783ade2-5aec-44e6-b0c0-7ffadf644661 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.492 NewBaseBdev 00:08:59.492 [2024-12-07 05:36:32.740809] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:59.492 [2024-12-07 05:36:32.740984] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:59.492 [2024-12-07 05:36:32.740995] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:59.492 [2024-12-07 05:36:32.741238] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:59.492 [2024-12-07 05:36:32.741345] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:59.492 [2024-12-07 05:36:32.741358] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:59.492 [2024-12-07 05:36:32.741469] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.492 [ 00:08:59.492 { 00:08:59.492 "name": "NewBaseBdev", 00:08:59.492 "aliases": [ 00:08:59.492 "3783ade2-5aec-44e6-b0c0-7ffadf644661" 00:08:59.492 ], 00:08:59.492 "product_name": "Malloc disk", 00:08:59.492 "block_size": 512, 00:08:59.492 "num_blocks": 65536, 00:08:59.492 "uuid": "3783ade2-5aec-44e6-b0c0-7ffadf644661", 00:08:59.492 "assigned_rate_limits": { 00:08:59.492 "rw_ios_per_sec": 0, 00:08:59.492 "rw_mbytes_per_sec": 0, 00:08:59.492 "r_mbytes_per_sec": 0, 00:08:59.492 "w_mbytes_per_sec": 0 00:08:59.492 }, 00:08:59.492 "claimed": true, 00:08:59.492 "claim_type": "exclusive_write", 00:08:59.492 "zoned": false, 00:08:59.492 "supported_io_types": { 00:08:59.492 "read": true, 00:08:59.492 "write": true, 00:08:59.492 "unmap": true, 00:08:59.492 "flush": true, 00:08:59.492 "reset": true, 00:08:59.492 "nvme_admin": false, 00:08:59.492 "nvme_io": false, 00:08:59.492 "nvme_io_md": false, 00:08:59.492 "write_zeroes": true, 00:08:59.492 "zcopy": true, 00:08:59.492 "get_zone_info": false, 00:08:59.492 "zone_management": false, 00:08:59.492 "zone_append": false, 00:08:59.492 "compare": false, 00:08:59.492 "compare_and_write": false, 00:08:59.492 "abort": true, 00:08:59.492 "seek_hole": false, 00:08:59.492 "seek_data": false, 00:08:59.492 "copy": true, 00:08:59.492 "nvme_iov_md": false 00:08:59.492 }, 00:08:59.492 "memory_domains": [ 00:08:59.492 { 00:08:59.492 "dma_device_id": "system", 00:08:59.492 "dma_device_type": 1 00:08:59.492 }, 00:08:59.492 { 00:08:59.492 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.492 "dma_device_type": 2 00:08:59.492 } 00:08:59.492 ], 00:08:59.492 "driver_specific": {} 00:08:59.492 } 00:08:59.492 ] 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:59.492 "name": "Existed_Raid", 00:08:59.492 "uuid": "6fc5de2c-a9ea-4ce2-b773-51d0970a1cb5", 00:08:59.492 "strip_size_kb": 0, 00:08:59.492 "state": "online", 00:08:59.492 "raid_level": "raid1", 00:08:59.492 "superblock": true, 00:08:59.492 "num_base_bdevs": 3, 00:08:59.492 "num_base_bdevs_discovered": 3, 00:08:59.492 "num_base_bdevs_operational": 3, 00:08:59.492 "base_bdevs_list": [ 00:08:59.492 { 00:08:59.492 "name": "NewBaseBdev", 00:08:59.492 "uuid": "3783ade2-5aec-44e6-b0c0-7ffadf644661", 00:08:59.492 "is_configured": true, 00:08:59.492 "data_offset": 2048, 00:08:59.492 "data_size": 63488 00:08:59.492 }, 00:08:59.492 { 00:08:59.492 "name": "BaseBdev2", 00:08:59.492 "uuid": "1cc627b0-7767-4feb-9760-b3dbc3be7cfd", 00:08:59.492 "is_configured": true, 00:08:59.492 "data_offset": 2048, 00:08:59.492 "data_size": 63488 00:08:59.492 }, 00:08:59.492 { 00:08:59.492 "name": "BaseBdev3", 00:08:59.492 "uuid": "f72a6bcb-63f4-4b4f-8d60-73d0a1b98de7", 00:08:59.492 "is_configured": true, 00:08:59.492 "data_offset": 2048, 00:08:59.492 "data_size": 63488 00:08:59.492 } 00:08:59.492 ] 00:08:59.492 }' 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:59.492 05:36:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.058 [2024-12-07 05:36:33.200371] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:00.058 "name": "Existed_Raid", 00:09:00.058 "aliases": [ 00:09:00.058 "6fc5de2c-a9ea-4ce2-b773-51d0970a1cb5" 00:09:00.058 ], 00:09:00.058 "product_name": "Raid Volume", 00:09:00.058 "block_size": 512, 00:09:00.058 "num_blocks": 63488, 00:09:00.058 "uuid": "6fc5de2c-a9ea-4ce2-b773-51d0970a1cb5", 00:09:00.058 "assigned_rate_limits": { 00:09:00.058 "rw_ios_per_sec": 0, 00:09:00.058 "rw_mbytes_per_sec": 0, 00:09:00.058 "r_mbytes_per_sec": 0, 00:09:00.058 "w_mbytes_per_sec": 0 00:09:00.058 }, 00:09:00.058 "claimed": false, 00:09:00.058 "zoned": false, 00:09:00.058 "supported_io_types": { 00:09:00.058 "read": true, 00:09:00.058 "write": true, 00:09:00.058 "unmap": false, 00:09:00.058 "flush": false, 00:09:00.058 "reset": true, 00:09:00.058 "nvme_admin": false, 00:09:00.058 "nvme_io": false, 00:09:00.058 "nvme_io_md": false, 00:09:00.058 "write_zeroes": true, 00:09:00.058 "zcopy": false, 00:09:00.058 "get_zone_info": false, 00:09:00.058 "zone_management": false, 00:09:00.058 "zone_append": false, 00:09:00.058 "compare": false, 00:09:00.058 "compare_and_write": false, 00:09:00.058 "abort": false, 00:09:00.058 "seek_hole": false, 00:09:00.058 "seek_data": false, 00:09:00.058 "copy": false, 00:09:00.058 "nvme_iov_md": false 00:09:00.058 }, 00:09:00.058 "memory_domains": [ 00:09:00.058 { 00:09:00.058 "dma_device_id": "system", 00:09:00.058 "dma_device_type": 1 00:09:00.058 }, 00:09:00.058 { 00:09:00.058 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.058 "dma_device_type": 2 00:09:00.058 }, 00:09:00.058 { 00:09:00.058 "dma_device_id": "system", 00:09:00.058 "dma_device_type": 1 00:09:00.058 }, 00:09:00.058 { 00:09:00.058 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.058 "dma_device_type": 2 00:09:00.058 }, 00:09:00.058 { 00:09:00.058 "dma_device_id": "system", 00:09:00.058 "dma_device_type": 1 00:09:00.058 }, 00:09:00.058 { 00:09:00.058 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.058 "dma_device_type": 2 00:09:00.058 } 00:09:00.058 ], 00:09:00.058 "driver_specific": { 00:09:00.058 "raid": { 00:09:00.058 "uuid": "6fc5de2c-a9ea-4ce2-b773-51d0970a1cb5", 00:09:00.058 "strip_size_kb": 0, 00:09:00.058 "state": "online", 00:09:00.058 "raid_level": "raid1", 00:09:00.058 "superblock": true, 00:09:00.058 "num_base_bdevs": 3, 00:09:00.058 "num_base_bdevs_discovered": 3, 00:09:00.058 "num_base_bdevs_operational": 3, 00:09:00.058 "base_bdevs_list": [ 00:09:00.058 { 00:09:00.058 "name": "NewBaseBdev", 00:09:00.058 "uuid": "3783ade2-5aec-44e6-b0c0-7ffadf644661", 00:09:00.058 "is_configured": true, 00:09:00.058 "data_offset": 2048, 00:09:00.058 "data_size": 63488 00:09:00.058 }, 00:09:00.058 { 00:09:00.058 "name": "BaseBdev2", 00:09:00.058 "uuid": "1cc627b0-7767-4feb-9760-b3dbc3be7cfd", 00:09:00.058 "is_configured": true, 00:09:00.058 "data_offset": 2048, 00:09:00.058 "data_size": 63488 00:09:00.058 }, 00:09:00.058 { 00:09:00.058 "name": "BaseBdev3", 00:09:00.058 "uuid": "f72a6bcb-63f4-4b4f-8d60-73d0a1b98de7", 00:09:00.058 "is_configured": true, 00:09:00.058 "data_offset": 2048, 00:09:00.058 "data_size": 63488 00:09:00.058 } 00:09:00.058 ] 00:09:00.058 } 00:09:00.058 } 00:09:00.058 }' 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:00.058 BaseBdev2 00:09:00.058 BaseBdev3' 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.058 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.316 [2024-12-07 05:36:33.435692] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:00.316 [2024-12-07 05:36:33.435768] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:00.316 [2024-12-07 05:36:33.435876] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:00.316 [2024-12-07 05:36:33.436160] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:00.316 [2024-12-07 05:36:33.436212] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 78714 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 78714 ']' 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 78714 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78714 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78714' 00:09:00.316 killing process with pid 78714 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 78714 00:09:00.316 [2024-12-07 05:36:33.479641] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:00.316 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 78714 00:09:00.316 [2024-12-07 05:36:33.510577] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:00.575 ************************************ 00:09:00.575 END TEST raid_state_function_test_sb 00:09:00.575 ************************************ 00:09:00.575 05:36:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:00.575 00:09:00.575 real 0m8.863s 00:09:00.575 user 0m15.216s 00:09:00.575 sys 0m1.768s 00:09:00.575 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:00.575 05:36:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.575 05:36:33 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:09:00.575 05:36:33 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:00.575 05:36:33 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:00.575 05:36:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:00.575 ************************************ 00:09:00.575 START TEST raid_superblock_test 00:09:00.575 ************************************ 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 3 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=79317 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 79317 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 79317 ']' 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:00.575 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:00.575 05:36:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.575 [2024-12-07 05:36:33.882295] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:09:00.575 [2024-12-07 05:36:33.882516] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79317 ] 00:09:00.833 [2024-12-07 05:36:34.014398] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:00.833 [2024-12-07 05:36:34.040164] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:00.833 [2024-12-07 05:36:34.082927] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:00.833 [2024-12-07 05:36:34.083043] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.409 malloc1 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.409 [2024-12-07 05:36:34.762307] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:01.409 [2024-12-07 05:36:34.762376] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:01.409 [2024-12-07 05:36:34.762395] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:01.409 [2024-12-07 05:36:34.762409] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:01.409 [2024-12-07 05:36:34.764559] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:01.409 [2024-12-07 05:36:34.764672] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:01.409 pt1 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.409 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.669 malloc2 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.669 [2024-12-07 05:36:34.790885] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:01.669 [2024-12-07 05:36:34.790999] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:01.669 [2024-12-07 05:36:34.791060] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:01.669 [2024-12-07 05:36:34.791093] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:01.669 [2024-12-07 05:36:34.793177] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:01.669 [2024-12-07 05:36:34.793245] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:01.669 pt2 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.669 malloc3 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.669 [2024-12-07 05:36:34.823405] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:01.669 [2024-12-07 05:36:34.823504] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:01.669 [2024-12-07 05:36:34.823547] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:01.669 [2024-12-07 05:36:34.823580] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:01.669 [2024-12-07 05:36:34.825731] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:01.669 [2024-12-07 05:36:34.825800] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:01.669 pt3 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.669 [2024-12-07 05:36:34.835440] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:01.669 [2024-12-07 05:36:34.837353] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:01.669 [2024-12-07 05:36:34.837409] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:01.669 [2024-12-07 05:36:34.837550] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:01.669 [2024-12-07 05:36:34.837561] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:01.669 [2024-12-07 05:36:34.837826] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:01.669 [2024-12-07 05:36:34.837985] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:01.669 [2024-12-07 05:36:34.838002] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:01.669 [2024-12-07 05:36:34.838114] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.669 "name": "raid_bdev1", 00:09:01.669 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:01.669 "strip_size_kb": 0, 00:09:01.669 "state": "online", 00:09:01.669 "raid_level": "raid1", 00:09:01.669 "superblock": true, 00:09:01.669 "num_base_bdevs": 3, 00:09:01.669 "num_base_bdevs_discovered": 3, 00:09:01.669 "num_base_bdevs_operational": 3, 00:09:01.669 "base_bdevs_list": [ 00:09:01.669 { 00:09:01.669 "name": "pt1", 00:09:01.669 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:01.669 "is_configured": true, 00:09:01.669 "data_offset": 2048, 00:09:01.669 "data_size": 63488 00:09:01.669 }, 00:09:01.669 { 00:09:01.669 "name": "pt2", 00:09:01.669 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:01.669 "is_configured": true, 00:09:01.669 "data_offset": 2048, 00:09:01.669 "data_size": 63488 00:09:01.669 }, 00:09:01.669 { 00:09:01.669 "name": "pt3", 00:09:01.669 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:01.669 "is_configured": true, 00:09:01.669 "data_offset": 2048, 00:09:01.669 "data_size": 63488 00:09:01.669 } 00:09:01.669 ] 00:09:01.669 }' 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.669 05:36:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.929 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:01.929 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:01.929 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:01.929 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:01.929 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:01.929 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:01.929 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:01.929 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:01.929 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.929 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.929 [2024-12-07 05:36:35.275061] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:01.929 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.190 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:02.190 "name": "raid_bdev1", 00:09:02.190 "aliases": [ 00:09:02.190 "d1676e99-b5a1-4cc6-b999-8eb97962d806" 00:09:02.190 ], 00:09:02.190 "product_name": "Raid Volume", 00:09:02.190 "block_size": 512, 00:09:02.190 "num_blocks": 63488, 00:09:02.190 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:02.190 "assigned_rate_limits": { 00:09:02.190 "rw_ios_per_sec": 0, 00:09:02.190 "rw_mbytes_per_sec": 0, 00:09:02.190 "r_mbytes_per_sec": 0, 00:09:02.190 "w_mbytes_per_sec": 0 00:09:02.191 }, 00:09:02.191 "claimed": false, 00:09:02.191 "zoned": false, 00:09:02.191 "supported_io_types": { 00:09:02.191 "read": true, 00:09:02.191 "write": true, 00:09:02.191 "unmap": false, 00:09:02.191 "flush": false, 00:09:02.191 "reset": true, 00:09:02.191 "nvme_admin": false, 00:09:02.191 "nvme_io": false, 00:09:02.191 "nvme_io_md": false, 00:09:02.191 "write_zeroes": true, 00:09:02.191 "zcopy": false, 00:09:02.191 "get_zone_info": false, 00:09:02.191 "zone_management": false, 00:09:02.191 "zone_append": false, 00:09:02.191 "compare": false, 00:09:02.191 "compare_and_write": false, 00:09:02.191 "abort": false, 00:09:02.191 "seek_hole": false, 00:09:02.191 "seek_data": false, 00:09:02.191 "copy": false, 00:09:02.191 "nvme_iov_md": false 00:09:02.191 }, 00:09:02.191 "memory_domains": [ 00:09:02.191 { 00:09:02.191 "dma_device_id": "system", 00:09:02.191 "dma_device_type": 1 00:09:02.191 }, 00:09:02.191 { 00:09:02.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.191 "dma_device_type": 2 00:09:02.191 }, 00:09:02.191 { 00:09:02.191 "dma_device_id": "system", 00:09:02.191 "dma_device_type": 1 00:09:02.191 }, 00:09:02.191 { 00:09:02.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.191 "dma_device_type": 2 00:09:02.191 }, 00:09:02.191 { 00:09:02.191 "dma_device_id": "system", 00:09:02.191 "dma_device_type": 1 00:09:02.191 }, 00:09:02.191 { 00:09:02.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.191 "dma_device_type": 2 00:09:02.191 } 00:09:02.191 ], 00:09:02.191 "driver_specific": { 00:09:02.191 "raid": { 00:09:02.191 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:02.191 "strip_size_kb": 0, 00:09:02.191 "state": "online", 00:09:02.191 "raid_level": "raid1", 00:09:02.191 "superblock": true, 00:09:02.191 "num_base_bdevs": 3, 00:09:02.191 "num_base_bdevs_discovered": 3, 00:09:02.191 "num_base_bdevs_operational": 3, 00:09:02.191 "base_bdevs_list": [ 00:09:02.191 { 00:09:02.191 "name": "pt1", 00:09:02.191 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:02.191 "is_configured": true, 00:09:02.191 "data_offset": 2048, 00:09:02.191 "data_size": 63488 00:09:02.191 }, 00:09:02.191 { 00:09:02.191 "name": "pt2", 00:09:02.191 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:02.191 "is_configured": true, 00:09:02.191 "data_offset": 2048, 00:09:02.191 "data_size": 63488 00:09:02.191 }, 00:09:02.191 { 00:09:02.191 "name": "pt3", 00:09:02.191 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:02.191 "is_configured": true, 00:09:02.191 "data_offset": 2048, 00:09:02.191 "data_size": 63488 00:09:02.191 } 00:09:02.191 ] 00:09:02.191 } 00:09:02.191 } 00:09:02.191 }' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:02.191 pt2 00:09:02.191 pt3' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.191 [2024-12-07 05:36:35.534492] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:02.191 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=d1676e99-b5a1-4cc6-b999-8eb97962d806 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z d1676e99-b5a1-4cc6-b999-8eb97962d806 ']' 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.450 [2024-12-07 05:36:35.578157] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:02.450 [2024-12-07 05:36:35.578225] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:02.450 [2024-12-07 05:36:35.578328] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:02.450 [2024-12-07 05:36:35.578434] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:02.450 [2024-12-07 05:36:35.578506] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:02.450 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.451 [2024-12-07 05:36:35.721950] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:02.451 [2024-12-07 05:36:35.724062] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:02.451 [2024-12-07 05:36:35.724112] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:02.451 [2024-12-07 05:36:35.724165] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:02.451 [2024-12-07 05:36:35.724222] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:02.451 [2024-12-07 05:36:35.724244] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:02.451 [2024-12-07 05:36:35.724258] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:02.451 [2024-12-07 05:36:35.724270] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:02.451 request: 00:09:02.451 { 00:09:02.451 "name": "raid_bdev1", 00:09:02.451 "raid_level": "raid1", 00:09:02.451 "base_bdevs": [ 00:09:02.451 "malloc1", 00:09:02.451 "malloc2", 00:09:02.451 "malloc3" 00:09:02.451 ], 00:09:02.451 "superblock": false, 00:09:02.451 "method": "bdev_raid_create", 00:09:02.451 "req_id": 1 00:09:02.451 } 00:09:02.451 Got JSON-RPC error response 00:09:02.451 response: 00:09:02.451 { 00:09:02.451 "code": -17, 00:09:02.451 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:02.451 } 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.451 [2024-12-07 05:36:35.785804] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:02.451 [2024-12-07 05:36:35.785900] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:02.451 [2024-12-07 05:36:35.785943] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:02.451 [2024-12-07 05:36:35.785975] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:02.451 [2024-12-07 05:36:35.788190] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:02.451 [2024-12-07 05:36:35.788264] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:02.451 [2024-12-07 05:36:35.788360] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:02.451 [2024-12-07 05:36:35.788426] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:02.451 pt1 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.451 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.709 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.709 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.709 "name": "raid_bdev1", 00:09:02.709 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:02.709 "strip_size_kb": 0, 00:09:02.709 "state": "configuring", 00:09:02.709 "raid_level": "raid1", 00:09:02.709 "superblock": true, 00:09:02.709 "num_base_bdevs": 3, 00:09:02.709 "num_base_bdevs_discovered": 1, 00:09:02.709 "num_base_bdevs_operational": 3, 00:09:02.709 "base_bdevs_list": [ 00:09:02.709 { 00:09:02.709 "name": "pt1", 00:09:02.709 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:02.709 "is_configured": true, 00:09:02.709 "data_offset": 2048, 00:09:02.709 "data_size": 63488 00:09:02.709 }, 00:09:02.709 { 00:09:02.709 "name": null, 00:09:02.709 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:02.709 "is_configured": false, 00:09:02.709 "data_offset": 2048, 00:09:02.709 "data_size": 63488 00:09:02.709 }, 00:09:02.709 { 00:09:02.709 "name": null, 00:09:02.710 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:02.710 "is_configured": false, 00:09:02.710 "data_offset": 2048, 00:09:02.710 "data_size": 63488 00:09:02.710 } 00:09:02.710 ] 00:09:02.710 }' 00:09:02.710 05:36:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.710 05:36:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.969 [2024-12-07 05:36:36.245030] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:02.969 [2024-12-07 05:36:36.245100] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:02.969 [2024-12-07 05:36:36.245137] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:09:02.969 [2024-12-07 05:36:36.245150] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:02.969 [2024-12-07 05:36:36.245555] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:02.969 [2024-12-07 05:36:36.245575] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:02.969 [2024-12-07 05:36:36.245669] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:02.969 [2024-12-07 05:36:36.245705] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:02.969 pt2 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.969 [2024-12-07 05:36:36.253015] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.969 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.969 "name": "raid_bdev1", 00:09:02.969 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:02.969 "strip_size_kb": 0, 00:09:02.969 "state": "configuring", 00:09:02.969 "raid_level": "raid1", 00:09:02.969 "superblock": true, 00:09:02.969 "num_base_bdevs": 3, 00:09:02.969 "num_base_bdevs_discovered": 1, 00:09:02.969 "num_base_bdevs_operational": 3, 00:09:02.969 "base_bdevs_list": [ 00:09:02.969 { 00:09:02.969 "name": "pt1", 00:09:02.969 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:02.969 "is_configured": true, 00:09:02.969 "data_offset": 2048, 00:09:02.969 "data_size": 63488 00:09:02.969 }, 00:09:02.969 { 00:09:02.969 "name": null, 00:09:02.969 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:02.969 "is_configured": false, 00:09:02.970 "data_offset": 0, 00:09:02.970 "data_size": 63488 00:09:02.970 }, 00:09:02.970 { 00:09:02.970 "name": null, 00:09:02.970 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:02.970 "is_configured": false, 00:09:02.970 "data_offset": 2048, 00:09:02.970 "data_size": 63488 00:09:02.970 } 00:09:02.970 ] 00:09:02.970 }' 00:09:02.970 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.970 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.538 [2024-12-07 05:36:36.716235] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:03.538 [2024-12-07 05:36:36.716354] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:03.538 [2024-12-07 05:36:36.716415] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:03.538 [2024-12-07 05:36:36.716445] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:03.538 [2024-12-07 05:36:36.716898] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:03.538 [2024-12-07 05:36:36.716956] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:03.538 [2024-12-07 05:36:36.717060] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:03.538 [2024-12-07 05:36:36.717110] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:03.538 pt2 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.538 [2024-12-07 05:36:36.728199] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:03.538 [2024-12-07 05:36:36.728276] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:03.538 [2024-12-07 05:36:36.728335] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:03.538 [2024-12-07 05:36:36.728364] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:03.538 [2024-12-07 05:36:36.728725] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:03.538 [2024-12-07 05:36:36.728777] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:03.538 [2024-12-07 05:36:36.728868] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:03.538 [2024-12-07 05:36:36.728911] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:03.538 [2024-12-07 05:36:36.729039] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:03.538 [2024-12-07 05:36:36.729078] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:03.538 [2024-12-07 05:36:36.729351] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:03.538 [2024-12-07 05:36:36.729499] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:03.538 [2024-12-07 05:36:36.729540] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:03.538 [2024-12-07 05:36:36.729685] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:03.538 pt3 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:03.538 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.539 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:03.539 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.539 "name": "raid_bdev1", 00:09:03.539 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:03.539 "strip_size_kb": 0, 00:09:03.539 "state": "online", 00:09:03.539 "raid_level": "raid1", 00:09:03.539 "superblock": true, 00:09:03.539 "num_base_bdevs": 3, 00:09:03.539 "num_base_bdevs_discovered": 3, 00:09:03.539 "num_base_bdevs_operational": 3, 00:09:03.539 "base_bdevs_list": [ 00:09:03.539 { 00:09:03.539 "name": "pt1", 00:09:03.539 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:03.539 "is_configured": true, 00:09:03.539 "data_offset": 2048, 00:09:03.539 "data_size": 63488 00:09:03.539 }, 00:09:03.539 { 00:09:03.539 "name": "pt2", 00:09:03.539 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:03.539 "is_configured": true, 00:09:03.539 "data_offset": 2048, 00:09:03.539 "data_size": 63488 00:09:03.539 }, 00:09:03.539 { 00:09:03.539 "name": "pt3", 00:09:03.539 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:03.539 "is_configured": true, 00:09:03.539 "data_offset": 2048, 00:09:03.539 "data_size": 63488 00:09:03.539 } 00:09:03.539 ] 00:09:03.539 }' 00:09:03.539 05:36:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.539 05:36:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.107 [2024-12-07 05:36:37.215725] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:04.107 "name": "raid_bdev1", 00:09:04.107 "aliases": [ 00:09:04.107 "d1676e99-b5a1-4cc6-b999-8eb97962d806" 00:09:04.107 ], 00:09:04.107 "product_name": "Raid Volume", 00:09:04.107 "block_size": 512, 00:09:04.107 "num_blocks": 63488, 00:09:04.107 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:04.107 "assigned_rate_limits": { 00:09:04.107 "rw_ios_per_sec": 0, 00:09:04.107 "rw_mbytes_per_sec": 0, 00:09:04.107 "r_mbytes_per_sec": 0, 00:09:04.107 "w_mbytes_per_sec": 0 00:09:04.107 }, 00:09:04.107 "claimed": false, 00:09:04.107 "zoned": false, 00:09:04.107 "supported_io_types": { 00:09:04.107 "read": true, 00:09:04.107 "write": true, 00:09:04.107 "unmap": false, 00:09:04.107 "flush": false, 00:09:04.107 "reset": true, 00:09:04.107 "nvme_admin": false, 00:09:04.107 "nvme_io": false, 00:09:04.107 "nvme_io_md": false, 00:09:04.107 "write_zeroes": true, 00:09:04.107 "zcopy": false, 00:09:04.107 "get_zone_info": false, 00:09:04.107 "zone_management": false, 00:09:04.107 "zone_append": false, 00:09:04.107 "compare": false, 00:09:04.107 "compare_and_write": false, 00:09:04.107 "abort": false, 00:09:04.107 "seek_hole": false, 00:09:04.107 "seek_data": false, 00:09:04.107 "copy": false, 00:09:04.107 "nvme_iov_md": false 00:09:04.107 }, 00:09:04.107 "memory_domains": [ 00:09:04.107 { 00:09:04.107 "dma_device_id": "system", 00:09:04.107 "dma_device_type": 1 00:09:04.107 }, 00:09:04.107 { 00:09:04.107 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:04.107 "dma_device_type": 2 00:09:04.107 }, 00:09:04.107 { 00:09:04.107 "dma_device_id": "system", 00:09:04.107 "dma_device_type": 1 00:09:04.107 }, 00:09:04.107 { 00:09:04.107 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:04.107 "dma_device_type": 2 00:09:04.107 }, 00:09:04.107 { 00:09:04.107 "dma_device_id": "system", 00:09:04.107 "dma_device_type": 1 00:09:04.107 }, 00:09:04.107 { 00:09:04.107 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:04.107 "dma_device_type": 2 00:09:04.107 } 00:09:04.107 ], 00:09:04.107 "driver_specific": { 00:09:04.107 "raid": { 00:09:04.107 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:04.107 "strip_size_kb": 0, 00:09:04.107 "state": "online", 00:09:04.107 "raid_level": "raid1", 00:09:04.107 "superblock": true, 00:09:04.107 "num_base_bdevs": 3, 00:09:04.107 "num_base_bdevs_discovered": 3, 00:09:04.107 "num_base_bdevs_operational": 3, 00:09:04.107 "base_bdevs_list": [ 00:09:04.107 { 00:09:04.107 "name": "pt1", 00:09:04.107 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:04.107 "is_configured": true, 00:09:04.107 "data_offset": 2048, 00:09:04.107 "data_size": 63488 00:09:04.107 }, 00:09:04.107 { 00:09:04.107 "name": "pt2", 00:09:04.107 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:04.107 "is_configured": true, 00:09:04.107 "data_offset": 2048, 00:09:04.107 "data_size": 63488 00:09:04.107 }, 00:09:04.107 { 00:09:04.107 "name": "pt3", 00:09:04.107 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:04.107 "is_configured": true, 00:09:04.107 "data_offset": 2048, 00:09:04.107 "data_size": 63488 00:09:04.107 } 00:09:04.107 ] 00:09:04.107 } 00:09:04.107 } 00:09:04.107 }' 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:04.107 pt2 00:09:04.107 pt3' 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:04.107 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.366 [2024-12-07 05:36:37.487195] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' d1676e99-b5a1-4cc6-b999-8eb97962d806 '!=' d1676e99-b5a1-4cc6-b999-8eb97962d806 ']' 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.366 [2024-12-07 05:36:37.530915] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.366 "name": "raid_bdev1", 00:09:04.366 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:04.366 "strip_size_kb": 0, 00:09:04.366 "state": "online", 00:09:04.366 "raid_level": "raid1", 00:09:04.366 "superblock": true, 00:09:04.366 "num_base_bdevs": 3, 00:09:04.366 "num_base_bdevs_discovered": 2, 00:09:04.366 "num_base_bdevs_operational": 2, 00:09:04.366 "base_bdevs_list": [ 00:09:04.366 { 00:09:04.366 "name": null, 00:09:04.366 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:04.366 "is_configured": false, 00:09:04.366 "data_offset": 0, 00:09:04.366 "data_size": 63488 00:09:04.366 }, 00:09:04.366 { 00:09:04.366 "name": "pt2", 00:09:04.366 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:04.366 "is_configured": true, 00:09:04.366 "data_offset": 2048, 00:09:04.366 "data_size": 63488 00:09:04.366 }, 00:09:04.366 { 00:09:04.366 "name": "pt3", 00:09:04.366 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:04.366 "is_configured": true, 00:09:04.366 "data_offset": 2048, 00:09:04.366 "data_size": 63488 00:09:04.366 } 00:09:04.366 ] 00:09:04.366 }' 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.366 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.623 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:04.623 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.623 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.624 [2024-12-07 05:36:37.926256] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:04.624 [2024-12-07 05:36:37.926290] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:04.624 [2024-12-07 05:36:37.926372] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:04.624 [2024-12-07 05:36:37.926434] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:04.624 [2024-12-07 05:36:37.926444] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.624 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.882 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.882 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:04.882 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:04.882 05:36:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:09:04.882 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.882 05:36:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.882 [2024-12-07 05:36:38.010079] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:04.882 [2024-12-07 05:36:38.010129] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:04.882 [2024-12-07 05:36:38.010162] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:09:04.882 [2024-12-07 05:36:38.010170] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:04.882 [2024-12-07 05:36:38.012341] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:04.882 [2024-12-07 05:36:38.012418] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:04.882 [2024-12-07 05:36:38.012495] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:04.882 [2024-12-07 05:36:38.012527] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:04.882 pt2 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.882 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.882 "name": "raid_bdev1", 00:09:04.882 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:04.882 "strip_size_kb": 0, 00:09:04.882 "state": "configuring", 00:09:04.882 "raid_level": "raid1", 00:09:04.882 "superblock": true, 00:09:04.882 "num_base_bdevs": 3, 00:09:04.882 "num_base_bdevs_discovered": 1, 00:09:04.882 "num_base_bdevs_operational": 2, 00:09:04.882 "base_bdevs_list": [ 00:09:04.882 { 00:09:04.882 "name": null, 00:09:04.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:04.882 "is_configured": false, 00:09:04.882 "data_offset": 2048, 00:09:04.882 "data_size": 63488 00:09:04.882 }, 00:09:04.882 { 00:09:04.882 "name": "pt2", 00:09:04.882 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:04.882 "is_configured": true, 00:09:04.882 "data_offset": 2048, 00:09:04.883 "data_size": 63488 00:09:04.883 }, 00:09:04.883 { 00:09:04.883 "name": null, 00:09:04.883 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:04.883 "is_configured": false, 00:09:04.883 "data_offset": 2048, 00:09:04.883 "data_size": 63488 00:09:04.883 } 00:09:04.883 ] 00:09:04.883 }' 00:09:04.883 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.883 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.141 [2024-12-07 05:36:38.465330] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:05.141 [2024-12-07 05:36:38.465466] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:05.141 [2024-12-07 05:36:38.465520] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:05.141 [2024-12-07 05:36:38.465552] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:05.141 [2024-12-07 05:36:38.465984] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:05.141 [2024-12-07 05:36:38.466042] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:05.141 [2024-12-07 05:36:38.466147] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:05.141 [2024-12-07 05:36:38.466198] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:05.141 [2024-12-07 05:36:38.466321] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:05.141 [2024-12-07 05:36:38.466361] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:05.141 [2024-12-07 05:36:38.466653] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:05.141 [2024-12-07 05:36:38.466820] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:05.141 [2024-12-07 05:36:38.466864] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:05.141 [2024-12-07 05:36:38.467006] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:05.141 pt3 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:05.141 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.401 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.401 "name": "raid_bdev1", 00:09:05.401 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:05.401 "strip_size_kb": 0, 00:09:05.401 "state": "online", 00:09:05.401 "raid_level": "raid1", 00:09:05.401 "superblock": true, 00:09:05.401 "num_base_bdevs": 3, 00:09:05.401 "num_base_bdevs_discovered": 2, 00:09:05.401 "num_base_bdevs_operational": 2, 00:09:05.401 "base_bdevs_list": [ 00:09:05.401 { 00:09:05.401 "name": null, 00:09:05.401 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:05.401 "is_configured": false, 00:09:05.401 "data_offset": 2048, 00:09:05.401 "data_size": 63488 00:09:05.401 }, 00:09:05.401 { 00:09:05.401 "name": "pt2", 00:09:05.401 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:05.401 "is_configured": true, 00:09:05.401 "data_offset": 2048, 00:09:05.401 "data_size": 63488 00:09:05.401 }, 00:09:05.401 { 00:09:05.401 "name": "pt3", 00:09:05.401 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:05.401 "is_configured": true, 00:09:05.401 "data_offset": 2048, 00:09:05.401 "data_size": 63488 00:09:05.401 } 00:09:05.401 ] 00:09:05.401 }' 00:09:05.401 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.401 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.661 [2024-12-07 05:36:38.904585] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:05.661 [2024-12-07 05:36:38.904696] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:05.661 [2024-12-07 05:36:38.904782] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:05.661 [2024-12-07 05:36:38.904842] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:05.661 [2024-12-07 05:36:38.904853] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.661 [2024-12-07 05:36:38.988430] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:05.661 [2024-12-07 05:36:38.988488] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:05.661 [2024-12-07 05:36:38.988508] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:05.661 [2024-12-07 05:36:38.988519] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:05.661 [2024-12-07 05:36:38.990662] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:05.661 [2024-12-07 05:36:38.990739] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:05.661 [2024-12-07 05:36:38.990818] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:05.661 [2024-12-07 05:36:38.990868] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:05.661 [2024-12-07 05:36:38.990987] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:09:05.661 [2024-12-07 05:36:38.991003] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:05.661 [2024-12-07 05:36:38.991019] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:09:05.661 [2024-12-07 05:36:38.991063] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:05.661 pt1 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.661 05:36:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:05.661 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.661 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.661 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.920 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.920 "name": "raid_bdev1", 00:09:05.920 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:05.920 "strip_size_kb": 0, 00:09:05.920 "state": "configuring", 00:09:05.920 "raid_level": "raid1", 00:09:05.920 "superblock": true, 00:09:05.920 "num_base_bdevs": 3, 00:09:05.920 "num_base_bdevs_discovered": 1, 00:09:05.920 "num_base_bdevs_operational": 2, 00:09:05.920 "base_bdevs_list": [ 00:09:05.920 { 00:09:05.920 "name": null, 00:09:05.920 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:05.920 "is_configured": false, 00:09:05.920 "data_offset": 2048, 00:09:05.920 "data_size": 63488 00:09:05.920 }, 00:09:05.920 { 00:09:05.920 "name": "pt2", 00:09:05.920 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:05.920 "is_configured": true, 00:09:05.920 "data_offset": 2048, 00:09:05.920 "data_size": 63488 00:09:05.920 }, 00:09:05.920 { 00:09:05.920 "name": null, 00:09:05.920 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:05.920 "is_configured": false, 00:09:05.921 "data_offset": 2048, 00:09:05.921 "data_size": 63488 00:09:05.921 } 00:09:05.921 ] 00:09:05.921 }' 00:09:05.921 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.921 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.180 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:06.180 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:09:06.180 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.180 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.180 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.180 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:09:06.180 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:06.180 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.180 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.180 [2024-12-07 05:36:39.431689] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:06.180 [2024-12-07 05:36:39.431796] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:06.180 [2024-12-07 05:36:39.431839] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:09:06.180 [2024-12-07 05:36:39.431875] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:06.180 [2024-12-07 05:36:39.432307] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:06.180 [2024-12-07 05:36:39.432366] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:06.180 [2024-12-07 05:36:39.432460] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:06.180 [2024-12-07 05:36:39.432530] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:06.180 [2024-12-07 05:36:39.432656] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:09:06.180 [2024-12-07 05:36:39.432697] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:06.180 [2024-12-07 05:36:39.432955] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:09:06.180 [2024-12-07 05:36:39.433116] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:09:06.180 [2024-12-07 05:36:39.433154] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:09:06.181 [2024-12-07 05:36:39.433301] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:06.181 pt3 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:06.181 "name": "raid_bdev1", 00:09:06.181 "uuid": "d1676e99-b5a1-4cc6-b999-8eb97962d806", 00:09:06.181 "strip_size_kb": 0, 00:09:06.181 "state": "online", 00:09:06.181 "raid_level": "raid1", 00:09:06.181 "superblock": true, 00:09:06.181 "num_base_bdevs": 3, 00:09:06.181 "num_base_bdevs_discovered": 2, 00:09:06.181 "num_base_bdevs_operational": 2, 00:09:06.181 "base_bdevs_list": [ 00:09:06.181 { 00:09:06.181 "name": null, 00:09:06.181 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:06.181 "is_configured": false, 00:09:06.181 "data_offset": 2048, 00:09:06.181 "data_size": 63488 00:09:06.181 }, 00:09:06.181 { 00:09:06.181 "name": "pt2", 00:09:06.181 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:06.181 "is_configured": true, 00:09:06.181 "data_offset": 2048, 00:09:06.181 "data_size": 63488 00:09:06.181 }, 00:09:06.181 { 00:09:06.181 "name": "pt3", 00:09:06.181 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:06.181 "is_configured": true, 00:09:06.181 "data_offset": 2048, 00:09:06.181 "data_size": 63488 00:09:06.181 } 00:09:06.181 ] 00:09:06.181 }' 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:06.181 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.749 [2024-12-07 05:36:39.943091] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' d1676e99-b5a1-4cc6-b999-8eb97962d806 '!=' d1676e99-b5a1-4cc6-b999-8eb97962d806 ']' 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 79317 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 79317 ']' 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 79317 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:06.749 05:36:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79317 00:09:06.749 killing process with pid 79317 00:09:06.749 05:36:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:06.749 05:36:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:06.749 05:36:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79317' 00:09:06.749 05:36:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 79317 00:09:06.749 [2024-12-07 05:36:40.030838] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:06.749 [2024-12-07 05:36:40.030929] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:06.749 05:36:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 79317 00:09:06.749 [2024-12-07 05:36:40.030997] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:06.749 [2024-12-07 05:36:40.031007] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:09:06.749 [2024-12-07 05:36:40.065373] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:07.008 05:36:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:07.008 00:09:07.008 real 0m6.465s 00:09:07.008 user 0m10.882s 00:09:07.008 sys 0m1.327s 00:09:07.008 ************************************ 00:09:07.008 END TEST raid_superblock_test 00:09:07.008 ************************************ 00:09:07.008 05:36:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:07.008 05:36:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.008 05:36:40 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:09:07.008 05:36:40 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:07.008 05:36:40 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:07.008 05:36:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:07.008 ************************************ 00:09:07.008 START TEST raid_read_error_test 00:09:07.008 ************************************ 00:09:07.008 05:36:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 read 00:09:07.008 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:07.008 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:07.008 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:07.008 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:07.008 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:07.008 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:07.008 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:07.008 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.gkOl0oDEJm 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=79752 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 79752 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 79752 ']' 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:07.009 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:07.009 05:36:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.268 [2024-12-07 05:36:40.431676] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:09:07.268 [2024-12-07 05:36:40.431884] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79752 ] 00:09:07.268 [2024-12-07 05:36:40.588686] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:07.268 [2024-12-07 05:36:40.615271] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:07.527 [2024-12-07 05:36:40.659973] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:07.527 [2024-12-07 05:36:40.660088] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.095 BaseBdev1_malloc 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.095 true 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.095 [2024-12-07 05:36:41.303912] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:08.095 [2024-12-07 05:36:41.303962] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.095 [2024-12-07 05:36:41.303980] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:08.095 [2024-12-07 05:36:41.303996] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.095 [2024-12-07 05:36:41.306121] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.095 [2024-12-07 05:36:41.306159] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:08.095 BaseBdev1 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.095 BaseBdev2_malloc 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.095 true 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.095 [2024-12-07 05:36:41.344509] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:08.095 [2024-12-07 05:36:41.344572] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.095 [2024-12-07 05:36:41.344605] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:08.095 [2024-12-07 05:36:41.344622] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.095 [2024-12-07 05:36:41.346974] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.095 [2024-12-07 05:36:41.347069] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:08.095 BaseBdev2 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.095 BaseBdev3_malloc 00:09:08.095 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.096 true 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.096 [2024-12-07 05:36:41.385079] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:08.096 [2024-12-07 05:36:41.385124] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.096 [2024-12-07 05:36:41.385143] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:08.096 [2024-12-07 05:36:41.385151] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.096 [2024-12-07 05:36:41.387284] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.096 [2024-12-07 05:36:41.387321] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:08.096 BaseBdev3 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.096 [2024-12-07 05:36:41.397120] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:08.096 [2024-12-07 05:36:41.399042] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:08.096 [2024-12-07 05:36:41.399119] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:08.096 [2024-12-07 05:36:41.399297] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:08.096 [2024-12-07 05:36:41.399313] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:08.096 [2024-12-07 05:36:41.399568] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:09:08.096 [2024-12-07 05:36:41.399755] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:08.096 [2024-12-07 05:36:41.399769] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:08.096 [2024-12-07 05:36:41.399904] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:08.096 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.355 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.355 "name": "raid_bdev1", 00:09:08.355 "uuid": "c3765043-bd63-45ea-9ea1-994e5e9a67d3", 00:09:08.355 "strip_size_kb": 0, 00:09:08.355 "state": "online", 00:09:08.355 "raid_level": "raid1", 00:09:08.355 "superblock": true, 00:09:08.355 "num_base_bdevs": 3, 00:09:08.355 "num_base_bdevs_discovered": 3, 00:09:08.355 "num_base_bdevs_operational": 3, 00:09:08.355 "base_bdevs_list": [ 00:09:08.355 { 00:09:08.355 "name": "BaseBdev1", 00:09:08.355 "uuid": "4d12e8f3-8d21-5187-bbc3-0cd1e296a041", 00:09:08.355 "is_configured": true, 00:09:08.355 "data_offset": 2048, 00:09:08.355 "data_size": 63488 00:09:08.355 }, 00:09:08.355 { 00:09:08.355 "name": "BaseBdev2", 00:09:08.355 "uuid": "95b1179e-c746-599c-8513-c75d42e6a2d8", 00:09:08.355 "is_configured": true, 00:09:08.355 "data_offset": 2048, 00:09:08.355 "data_size": 63488 00:09:08.355 }, 00:09:08.355 { 00:09:08.355 "name": "BaseBdev3", 00:09:08.355 "uuid": "8ae9ee6c-ed47-51c8-89ae-5d2a3add65b8", 00:09:08.355 "is_configured": true, 00:09:08.355 "data_offset": 2048, 00:09:08.355 "data_size": 63488 00:09:08.355 } 00:09:08.355 ] 00:09:08.355 }' 00:09:08.355 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.355 05:36:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.614 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:08.614 05:36:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:08.614 [2024-12-07 05:36:41.952587] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.553 05:36:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.813 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.813 "name": "raid_bdev1", 00:09:09.813 "uuid": "c3765043-bd63-45ea-9ea1-994e5e9a67d3", 00:09:09.813 "strip_size_kb": 0, 00:09:09.813 "state": "online", 00:09:09.813 "raid_level": "raid1", 00:09:09.813 "superblock": true, 00:09:09.813 "num_base_bdevs": 3, 00:09:09.813 "num_base_bdevs_discovered": 3, 00:09:09.813 "num_base_bdevs_operational": 3, 00:09:09.813 "base_bdevs_list": [ 00:09:09.813 { 00:09:09.813 "name": "BaseBdev1", 00:09:09.813 "uuid": "4d12e8f3-8d21-5187-bbc3-0cd1e296a041", 00:09:09.813 "is_configured": true, 00:09:09.813 "data_offset": 2048, 00:09:09.813 "data_size": 63488 00:09:09.813 }, 00:09:09.813 { 00:09:09.813 "name": "BaseBdev2", 00:09:09.813 "uuid": "95b1179e-c746-599c-8513-c75d42e6a2d8", 00:09:09.813 "is_configured": true, 00:09:09.813 "data_offset": 2048, 00:09:09.813 "data_size": 63488 00:09:09.813 }, 00:09:09.813 { 00:09:09.813 "name": "BaseBdev3", 00:09:09.813 "uuid": "8ae9ee6c-ed47-51c8-89ae-5d2a3add65b8", 00:09:09.813 "is_configured": true, 00:09:09.813 "data_offset": 2048, 00:09:09.813 "data_size": 63488 00:09:09.813 } 00:09:09.813 ] 00:09:09.813 }' 00:09:09.813 05:36:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.813 05:36:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.072 05:36:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:10.072 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.072 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.072 [2024-12-07 05:36:43.291377] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:10.072 [2024-12-07 05:36:43.291470] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:10.072 [2024-12-07 05:36:43.294181] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:10.072 [2024-12-07 05:36:43.294273] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:10.072 [2024-12-07 05:36:43.294401] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:10.072 [2024-12-07 05:36:43.294447] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:10.073 { 00:09:10.073 "results": [ 00:09:10.073 { 00:09:10.073 "job": "raid_bdev1", 00:09:10.073 "core_mask": "0x1", 00:09:10.073 "workload": "randrw", 00:09:10.073 "percentage": 50, 00:09:10.073 "status": "finished", 00:09:10.073 "queue_depth": 1, 00:09:10.073 "io_size": 131072, 00:09:10.073 "runtime": 1.339665, 00:09:10.073 "iops": 13799.71858636301, 00:09:10.073 "mibps": 1724.9648232953762, 00:09:10.073 "io_failed": 0, 00:09:10.073 "io_timeout": 0, 00:09:10.073 "avg_latency_us": 69.71681448287868, 00:09:10.073 "min_latency_us": 23.923144104803495, 00:09:10.073 "max_latency_us": 1502.46288209607 00:09:10.073 } 00:09:10.073 ], 00:09:10.073 "core_count": 1 00:09:10.073 } 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 79752 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 79752 ']' 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 79752 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79752 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79752' 00:09:10.073 killing process with pid 79752 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 79752 00:09:10.073 [2024-12-07 05:36:43.337659] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:10.073 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 79752 00:09:10.073 [2024-12-07 05:36:43.363068] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:10.332 05:36:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.gkOl0oDEJm 00:09:10.332 05:36:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:10.332 05:36:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:10.332 05:36:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:10.332 05:36:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:10.332 05:36:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:10.332 05:36:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:10.332 05:36:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:10.332 00:09:10.332 real 0m3.242s 00:09:10.332 user 0m4.120s 00:09:10.332 sys 0m0.517s 00:09:10.332 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:10.332 ************************************ 00:09:10.332 END TEST raid_read_error_test 00:09:10.332 ************************************ 00:09:10.332 05:36:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.332 05:36:43 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:09:10.332 05:36:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:10.332 05:36:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:10.332 05:36:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:10.332 ************************************ 00:09:10.332 START TEST raid_write_error_test 00:09:10.332 ************************************ 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 write 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.dFjeKbhZqs 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=79881 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 79881 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 79881 ']' 00:09:10.332 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:10.332 05:36:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.591 [2024-12-07 05:36:43.744666] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:09:10.591 [2024-12-07 05:36:43.744798] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79881 ] 00:09:10.591 [2024-12-07 05:36:43.900920] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:10.591 [2024-12-07 05:36:43.926349] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:10.850 [2024-12-07 05:36:43.968930] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:10.850 [2024-12-07 05:36:43.968963] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.419 BaseBdev1_malloc 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.419 true 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.419 [2024-12-07 05:36:44.612277] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:11.419 [2024-12-07 05:36:44.612388] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.419 [2024-12-07 05:36:44.612417] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:11.419 [2024-12-07 05:36:44.612426] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.419 [2024-12-07 05:36:44.614518] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.419 [2024-12-07 05:36:44.614551] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:11.419 BaseBdev1 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.419 BaseBdev2_malloc 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.419 true 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.419 [2024-12-07 05:36:44.652901] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:11.419 [2024-12-07 05:36:44.652953] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.419 [2024-12-07 05:36:44.652987] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:11.419 [2024-12-07 05:36:44.653004] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.419 [2024-12-07 05:36:44.655179] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.419 [2024-12-07 05:36:44.655218] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:11.419 BaseBdev2 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.419 BaseBdev3_malloc 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.419 true 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.419 [2024-12-07 05:36:44.694035] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:11.419 [2024-12-07 05:36:44.694085] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.419 [2024-12-07 05:36:44.694104] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:11.419 [2024-12-07 05:36:44.694114] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.419 [2024-12-07 05:36:44.696349] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.419 [2024-12-07 05:36:44.696385] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:11.419 BaseBdev3 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.419 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.419 [2024-12-07 05:36:44.706050] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:11.420 [2024-12-07 05:36:44.707951] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:11.420 [2024-12-07 05:36:44.708026] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:11.420 [2024-12-07 05:36:44.708213] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:11.420 [2024-12-07 05:36:44.708227] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:11.420 [2024-12-07 05:36:44.708471] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:09:11.420 [2024-12-07 05:36:44.708611] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:11.420 [2024-12-07 05:36:44.708620] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:11.420 [2024-12-07 05:36:44.708745] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.420 "name": "raid_bdev1", 00:09:11.420 "uuid": "39ec074b-51da-41f0-9fb6-daa95f605aad", 00:09:11.420 "strip_size_kb": 0, 00:09:11.420 "state": "online", 00:09:11.420 "raid_level": "raid1", 00:09:11.420 "superblock": true, 00:09:11.420 "num_base_bdevs": 3, 00:09:11.420 "num_base_bdevs_discovered": 3, 00:09:11.420 "num_base_bdevs_operational": 3, 00:09:11.420 "base_bdevs_list": [ 00:09:11.420 { 00:09:11.420 "name": "BaseBdev1", 00:09:11.420 "uuid": "587f0d89-7c3f-5bf3-b195-f2318909da78", 00:09:11.420 "is_configured": true, 00:09:11.420 "data_offset": 2048, 00:09:11.420 "data_size": 63488 00:09:11.420 }, 00:09:11.420 { 00:09:11.420 "name": "BaseBdev2", 00:09:11.420 "uuid": "194aa428-23eb-5c57-a7c8-83b4ee499cd2", 00:09:11.420 "is_configured": true, 00:09:11.420 "data_offset": 2048, 00:09:11.420 "data_size": 63488 00:09:11.420 }, 00:09:11.420 { 00:09:11.420 "name": "BaseBdev3", 00:09:11.420 "uuid": "36e3ecec-b134-5f60-b579-d3af76ae5955", 00:09:11.420 "is_configured": true, 00:09:11.420 "data_offset": 2048, 00:09:11.420 "data_size": 63488 00:09:11.420 } 00:09:11.420 ] 00:09:11.420 }' 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.420 05:36:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.988 05:36:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:11.988 05:36:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:11.988 [2024-12-07 05:36:45.265507] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.934 [2024-12-07 05:36:46.181962] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:09:12.934 [2024-12-07 05:36:46.182086] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:12.934 [2024-12-07 05:36:46.182364] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002d50 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.934 "name": "raid_bdev1", 00:09:12.934 "uuid": "39ec074b-51da-41f0-9fb6-daa95f605aad", 00:09:12.934 "strip_size_kb": 0, 00:09:12.934 "state": "online", 00:09:12.934 "raid_level": "raid1", 00:09:12.934 "superblock": true, 00:09:12.934 "num_base_bdevs": 3, 00:09:12.934 "num_base_bdevs_discovered": 2, 00:09:12.934 "num_base_bdevs_operational": 2, 00:09:12.934 "base_bdevs_list": [ 00:09:12.934 { 00:09:12.934 "name": null, 00:09:12.934 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:12.934 "is_configured": false, 00:09:12.934 "data_offset": 0, 00:09:12.934 "data_size": 63488 00:09:12.934 }, 00:09:12.934 { 00:09:12.934 "name": "BaseBdev2", 00:09:12.934 "uuid": "194aa428-23eb-5c57-a7c8-83b4ee499cd2", 00:09:12.934 "is_configured": true, 00:09:12.934 "data_offset": 2048, 00:09:12.934 "data_size": 63488 00:09:12.934 }, 00:09:12.934 { 00:09:12.934 "name": "BaseBdev3", 00:09:12.934 "uuid": "36e3ecec-b134-5f60-b579-d3af76ae5955", 00:09:12.934 "is_configured": true, 00:09:12.934 "data_offset": 2048, 00:09:12.934 "data_size": 63488 00:09:12.934 } 00:09:12.934 ] 00:09:12.934 }' 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.934 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.512 [2024-12-07 05:36:46.656214] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:13.512 [2024-12-07 05:36:46.656311] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:13.512 [2024-12-07 05:36:46.658819] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:13.512 [2024-12-07 05:36:46.658906] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:13.512 [2024-12-07 05:36:46.659018] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:13.512 [2024-12-07 05:36:46.659063] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:13.512 { 00:09:13.512 "results": [ 00:09:13.512 { 00:09:13.512 "job": "raid_bdev1", 00:09:13.512 "core_mask": "0x1", 00:09:13.512 "workload": "randrw", 00:09:13.512 "percentage": 50, 00:09:13.512 "status": "finished", 00:09:13.512 "queue_depth": 1, 00:09:13.512 "io_size": 131072, 00:09:13.512 "runtime": 1.391566, 00:09:13.512 "iops": 14970.184669645565, 00:09:13.512 "mibps": 1871.2730837056956, 00:09:13.512 "io_failed": 0, 00:09:13.512 "io_timeout": 0, 00:09:13.512 "avg_latency_us": 63.96905533307843, 00:09:13.512 "min_latency_us": 23.923144104803495, 00:09:13.512 "max_latency_us": 1631.2454148471616 00:09:13.512 } 00:09:13.512 ], 00:09:13.512 "core_count": 1 00:09:13.512 } 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 79881 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 79881 ']' 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 79881 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79881 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79881' 00:09:13.512 killing process with pid 79881 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 79881 00:09:13.512 [2024-12-07 05:36:46.704996] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:13.512 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 79881 00:09:13.512 [2024-12-07 05:36:46.730135] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:13.788 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.dFjeKbhZqs 00:09:13.788 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:13.788 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:13.788 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:13.788 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:13.788 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:13.788 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:13.788 05:36:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:13.788 00:09:13.788 real 0m3.311s 00:09:13.788 user 0m4.220s 00:09:13.788 sys 0m0.527s 00:09:13.788 ************************************ 00:09:13.788 END TEST raid_write_error_test 00:09:13.788 ************************************ 00:09:13.788 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:13.788 05:36:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.788 05:36:47 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:09:13.788 05:36:47 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:13.788 05:36:47 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:09:13.788 05:36:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:13.788 05:36:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:13.788 05:36:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:13.788 ************************************ 00:09:13.788 START TEST raid_state_function_test 00:09:13.788 ************************************ 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 false 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=80008 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80008' 00:09:13.788 Process raid pid: 80008 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 80008 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 80008 ']' 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:13.788 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:13.788 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.788 [2024-12-07 05:36:47.114431] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:09:13.788 [2024-12-07 05:36:47.114655] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:14.045 [2024-12-07 05:36:47.269950] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:14.045 [2024-12-07 05:36:47.295510] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:14.045 [2024-12-07 05:36:47.337512] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:14.045 [2024-12-07 05:36:47.337548] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:14.613 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:14.613 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:14.613 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:14.613 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.613 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.613 [2024-12-07 05:36:47.952015] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:14.613 [2024-12-07 05:36:47.952078] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:14.613 [2024-12-07 05:36:47.952088] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:14.613 [2024-12-07 05:36:47.952099] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:14.613 [2024-12-07 05:36:47.952105] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:14.613 [2024-12-07 05:36:47.952116] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:14.613 [2024-12-07 05:36:47.952122] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:14.613 [2024-12-07 05:36:47.952130] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:14.613 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.613 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.614 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.872 05:36:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.872 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.872 "name": "Existed_Raid", 00:09:14.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.872 "strip_size_kb": 64, 00:09:14.872 "state": "configuring", 00:09:14.872 "raid_level": "raid0", 00:09:14.872 "superblock": false, 00:09:14.872 "num_base_bdevs": 4, 00:09:14.872 "num_base_bdevs_discovered": 0, 00:09:14.872 "num_base_bdevs_operational": 4, 00:09:14.872 "base_bdevs_list": [ 00:09:14.872 { 00:09:14.872 "name": "BaseBdev1", 00:09:14.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.872 "is_configured": false, 00:09:14.872 "data_offset": 0, 00:09:14.872 "data_size": 0 00:09:14.872 }, 00:09:14.872 { 00:09:14.872 "name": "BaseBdev2", 00:09:14.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.872 "is_configured": false, 00:09:14.872 "data_offset": 0, 00:09:14.872 "data_size": 0 00:09:14.872 }, 00:09:14.872 { 00:09:14.872 "name": "BaseBdev3", 00:09:14.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.872 "is_configured": false, 00:09:14.872 "data_offset": 0, 00:09:14.872 "data_size": 0 00:09:14.872 }, 00:09:14.872 { 00:09:14.872 "name": "BaseBdev4", 00:09:14.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.872 "is_configured": false, 00:09:14.872 "data_offset": 0, 00:09:14.872 "data_size": 0 00:09:14.872 } 00:09:14.872 ] 00:09:14.872 }' 00:09:14.872 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.872 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.130 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:15.130 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.130 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.130 [2024-12-07 05:36:48.423087] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:15.130 [2024-12-07 05:36:48.423179] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:15.130 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.130 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.131 [2024-12-07 05:36:48.435078] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:15.131 [2024-12-07 05:36:48.435163] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:15.131 [2024-12-07 05:36:48.435194] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:15.131 [2024-12-07 05:36:48.435247] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:15.131 [2024-12-07 05:36:48.435273] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:15.131 [2024-12-07 05:36:48.435303] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:15.131 [2024-12-07 05:36:48.435329] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:15.131 [2024-12-07 05:36:48.435357] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.131 [2024-12-07 05:36:48.456004] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:15.131 BaseBdev1 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.131 [ 00:09:15.131 { 00:09:15.131 "name": "BaseBdev1", 00:09:15.131 "aliases": [ 00:09:15.131 "5fb922eb-04d1-4d4c-8bbe-db0f5917744f" 00:09:15.131 ], 00:09:15.131 "product_name": "Malloc disk", 00:09:15.131 "block_size": 512, 00:09:15.131 "num_blocks": 65536, 00:09:15.131 "uuid": "5fb922eb-04d1-4d4c-8bbe-db0f5917744f", 00:09:15.131 "assigned_rate_limits": { 00:09:15.131 "rw_ios_per_sec": 0, 00:09:15.131 "rw_mbytes_per_sec": 0, 00:09:15.131 "r_mbytes_per_sec": 0, 00:09:15.131 "w_mbytes_per_sec": 0 00:09:15.131 }, 00:09:15.131 "claimed": true, 00:09:15.131 "claim_type": "exclusive_write", 00:09:15.131 "zoned": false, 00:09:15.131 "supported_io_types": { 00:09:15.131 "read": true, 00:09:15.131 "write": true, 00:09:15.131 "unmap": true, 00:09:15.131 "flush": true, 00:09:15.131 "reset": true, 00:09:15.131 "nvme_admin": false, 00:09:15.131 "nvme_io": false, 00:09:15.131 "nvme_io_md": false, 00:09:15.131 "write_zeroes": true, 00:09:15.131 "zcopy": true, 00:09:15.131 "get_zone_info": false, 00:09:15.131 "zone_management": false, 00:09:15.131 "zone_append": false, 00:09:15.131 "compare": false, 00:09:15.131 "compare_and_write": false, 00:09:15.131 "abort": true, 00:09:15.131 "seek_hole": false, 00:09:15.131 "seek_data": false, 00:09:15.131 "copy": true, 00:09:15.131 "nvme_iov_md": false 00:09:15.131 }, 00:09:15.131 "memory_domains": [ 00:09:15.131 { 00:09:15.131 "dma_device_id": "system", 00:09:15.131 "dma_device_type": 1 00:09:15.131 }, 00:09:15.131 { 00:09:15.131 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:15.131 "dma_device_type": 2 00:09:15.131 } 00:09:15.131 ], 00:09:15.131 "driver_specific": {} 00:09:15.131 } 00:09:15.131 ] 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:15.131 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:15.390 "name": "Existed_Raid", 00:09:15.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:15.390 "strip_size_kb": 64, 00:09:15.390 "state": "configuring", 00:09:15.390 "raid_level": "raid0", 00:09:15.390 "superblock": false, 00:09:15.390 "num_base_bdevs": 4, 00:09:15.390 "num_base_bdevs_discovered": 1, 00:09:15.390 "num_base_bdevs_operational": 4, 00:09:15.390 "base_bdevs_list": [ 00:09:15.390 { 00:09:15.390 "name": "BaseBdev1", 00:09:15.390 "uuid": "5fb922eb-04d1-4d4c-8bbe-db0f5917744f", 00:09:15.390 "is_configured": true, 00:09:15.390 "data_offset": 0, 00:09:15.390 "data_size": 65536 00:09:15.390 }, 00:09:15.390 { 00:09:15.390 "name": "BaseBdev2", 00:09:15.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:15.390 "is_configured": false, 00:09:15.390 "data_offset": 0, 00:09:15.390 "data_size": 0 00:09:15.390 }, 00:09:15.390 { 00:09:15.390 "name": "BaseBdev3", 00:09:15.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:15.390 "is_configured": false, 00:09:15.390 "data_offset": 0, 00:09:15.390 "data_size": 0 00:09:15.390 }, 00:09:15.390 { 00:09:15.390 "name": "BaseBdev4", 00:09:15.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:15.390 "is_configured": false, 00:09:15.390 "data_offset": 0, 00:09:15.390 "data_size": 0 00:09:15.390 } 00:09:15.390 ] 00:09:15.390 }' 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:15.390 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.650 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:15.650 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.650 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.650 [2024-12-07 05:36:48.991196] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:15.650 [2024-12-07 05:36:48.991262] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:15.650 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.650 05:36:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:15.650 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.650 05:36:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.650 [2024-12-07 05:36:49.003210] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:15.650 [2024-12-07 05:36:49.005058] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:15.650 [2024-12-07 05:36:49.005160] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:15.650 [2024-12-07 05:36:49.005175] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:15.650 [2024-12-07 05:36:49.005184] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:15.650 [2024-12-07 05:36:49.005190] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:15.650 [2024-12-07 05:36:49.005198] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.650 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.909 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:15.909 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.909 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:15.909 "name": "Existed_Raid", 00:09:15.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:15.909 "strip_size_kb": 64, 00:09:15.909 "state": "configuring", 00:09:15.909 "raid_level": "raid0", 00:09:15.909 "superblock": false, 00:09:15.909 "num_base_bdevs": 4, 00:09:15.909 "num_base_bdevs_discovered": 1, 00:09:15.909 "num_base_bdevs_operational": 4, 00:09:15.909 "base_bdevs_list": [ 00:09:15.909 { 00:09:15.909 "name": "BaseBdev1", 00:09:15.909 "uuid": "5fb922eb-04d1-4d4c-8bbe-db0f5917744f", 00:09:15.909 "is_configured": true, 00:09:15.909 "data_offset": 0, 00:09:15.909 "data_size": 65536 00:09:15.909 }, 00:09:15.909 { 00:09:15.909 "name": "BaseBdev2", 00:09:15.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:15.909 "is_configured": false, 00:09:15.909 "data_offset": 0, 00:09:15.909 "data_size": 0 00:09:15.909 }, 00:09:15.909 { 00:09:15.909 "name": "BaseBdev3", 00:09:15.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:15.909 "is_configured": false, 00:09:15.909 "data_offset": 0, 00:09:15.909 "data_size": 0 00:09:15.909 }, 00:09:15.909 { 00:09:15.909 "name": "BaseBdev4", 00:09:15.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:15.909 "is_configured": false, 00:09:15.909 "data_offset": 0, 00:09:15.909 "data_size": 0 00:09:15.909 } 00:09:15.909 ] 00:09:15.909 }' 00:09:15.909 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:15.909 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.167 [2024-12-07 05:36:49.493252] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:16.167 BaseBdev2 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.167 [ 00:09:16.167 { 00:09:16.167 "name": "BaseBdev2", 00:09:16.167 "aliases": [ 00:09:16.167 "be996da4-bbeb-4920-ac63-e5cc4d61c8d3" 00:09:16.167 ], 00:09:16.167 "product_name": "Malloc disk", 00:09:16.167 "block_size": 512, 00:09:16.167 "num_blocks": 65536, 00:09:16.167 "uuid": "be996da4-bbeb-4920-ac63-e5cc4d61c8d3", 00:09:16.167 "assigned_rate_limits": { 00:09:16.167 "rw_ios_per_sec": 0, 00:09:16.167 "rw_mbytes_per_sec": 0, 00:09:16.167 "r_mbytes_per_sec": 0, 00:09:16.167 "w_mbytes_per_sec": 0 00:09:16.167 }, 00:09:16.167 "claimed": true, 00:09:16.167 "claim_type": "exclusive_write", 00:09:16.167 "zoned": false, 00:09:16.167 "supported_io_types": { 00:09:16.167 "read": true, 00:09:16.167 "write": true, 00:09:16.167 "unmap": true, 00:09:16.167 "flush": true, 00:09:16.167 "reset": true, 00:09:16.167 "nvme_admin": false, 00:09:16.167 "nvme_io": false, 00:09:16.167 "nvme_io_md": false, 00:09:16.167 "write_zeroes": true, 00:09:16.167 "zcopy": true, 00:09:16.167 "get_zone_info": false, 00:09:16.167 "zone_management": false, 00:09:16.167 "zone_append": false, 00:09:16.167 "compare": false, 00:09:16.167 "compare_and_write": false, 00:09:16.167 "abort": true, 00:09:16.167 "seek_hole": false, 00:09:16.167 "seek_data": false, 00:09:16.167 "copy": true, 00:09:16.167 "nvme_iov_md": false 00:09:16.167 }, 00:09:16.167 "memory_domains": [ 00:09:16.167 { 00:09:16.167 "dma_device_id": "system", 00:09:16.167 "dma_device_type": 1 00:09:16.167 }, 00:09:16.167 { 00:09:16.167 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:16.167 "dma_device_type": 2 00:09:16.167 } 00:09:16.167 ], 00:09:16.167 "driver_specific": {} 00:09:16.167 } 00:09:16.167 ] 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:16.167 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:16.426 "name": "Existed_Raid", 00:09:16.426 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.426 "strip_size_kb": 64, 00:09:16.426 "state": "configuring", 00:09:16.426 "raid_level": "raid0", 00:09:16.426 "superblock": false, 00:09:16.426 "num_base_bdevs": 4, 00:09:16.426 "num_base_bdevs_discovered": 2, 00:09:16.426 "num_base_bdevs_operational": 4, 00:09:16.426 "base_bdevs_list": [ 00:09:16.426 { 00:09:16.426 "name": "BaseBdev1", 00:09:16.426 "uuid": "5fb922eb-04d1-4d4c-8bbe-db0f5917744f", 00:09:16.426 "is_configured": true, 00:09:16.426 "data_offset": 0, 00:09:16.426 "data_size": 65536 00:09:16.426 }, 00:09:16.426 { 00:09:16.426 "name": "BaseBdev2", 00:09:16.426 "uuid": "be996da4-bbeb-4920-ac63-e5cc4d61c8d3", 00:09:16.426 "is_configured": true, 00:09:16.426 "data_offset": 0, 00:09:16.426 "data_size": 65536 00:09:16.426 }, 00:09:16.426 { 00:09:16.426 "name": "BaseBdev3", 00:09:16.426 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.426 "is_configured": false, 00:09:16.426 "data_offset": 0, 00:09:16.426 "data_size": 0 00:09:16.426 }, 00:09:16.426 { 00:09:16.426 "name": "BaseBdev4", 00:09:16.426 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.426 "is_configured": false, 00:09:16.426 "data_offset": 0, 00:09:16.426 "data_size": 0 00:09:16.426 } 00:09:16.426 ] 00:09:16.426 }' 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:16.426 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.685 [2024-12-07 05:36:49.994931] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:16.685 BaseBdev3 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.685 05:36:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.685 [ 00:09:16.685 { 00:09:16.685 "name": "BaseBdev3", 00:09:16.685 "aliases": [ 00:09:16.685 "e30d3b71-e051-48ce-b24a-fe1a4b7e76ab" 00:09:16.685 ], 00:09:16.685 "product_name": "Malloc disk", 00:09:16.685 "block_size": 512, 00:09:16.685 "num_blocks": 65536, 00:09:16.685 "uuid": "e30d3b71-e051-48ce-b24a-fe1a4b7e76ab", 00:09:16.685 "assigned_rate_limits": { 00:09:16.685 "rw_ios_per_sec": 0, 00:09:16.685 "rw_mbytes_per_sec": 0, 00:09:16.685 "r_mbytes_per_sec": 0, 00:09:16.685 "w_mbytes_per_sec": 0 00:09:16.685 }, 00:09:16.685 "claimed": true, 00:09:16.685 "claim_type": "exclusive_write", 00:09:16.685 "zoned": false, 00:09:16.685 "supported_io_types": { 00:09:16.685 "read": true, 00:09:16.685 "write": true, 00:09:16.685 "unmap": true, 00:09:16.685 "flush": true, 00:09:16.685 "reset": true, 00:09:16.685 "nvme_admin": false, 00:09:16.685 "nvme_io": false, 00:09:16.685 "nvme_io_md": false, 00:09:16.685 "write_zeroes": true, 00:09:16.685 "zcopy": true, 00:09:16.685 "get_zone_info": false, 00:09:16.685 "zone_management": false, 00:09:16.685 "zone_append": false, 00:09:16.685 "compare": false, 00:09:16.685 "compare_and_write": false, 00:09:16.685 "abort": true, 00:09:16.685 "seek_hole": false, 00:09:16.685 "seek_data": false, 00:09:16.685 "copy": true, 00:09:16.685 "nvme_iov_md": false 00:09:16.685 }, 00:09:16.685 "memory_domains": [ 00:09:16.685 { 00:09:16.685 "dma_device_id": "system", 00:09:16.685 "dma_device_type": 1 00:09:16.685 }, 00:09:16.685 { 00:09:16.685 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:16.685 "dma_device_type": 2 00:09:16.685 } 00:09:16.685 ], 00:09:16.685 "driver_specific": {} 00:09:16.685 } 00:09:16.685 ] 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:16.685 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.686 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.943 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.943 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:16.943 "name": "Existed_Raid", 00:09:16.943 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.943 "strip_size_kb": 64, 00:09:16.943 "state": "configuring", 00:09:16.943 "raid_level": "raid0", 00:09:16.943 "superblock": false, 00:09:16.943 "num_base_bdevs": 4, 00:09:16.943 "num_base_bdevs_discovered": 3, 00:09:16.943 "num_base_bdevs_operational": 4, 00:09:16.943 "base_bdevs_list": [ 00:09:16.943 { 00:09:16.943 "name": "BaseBdev1", 00:09:16.943 "uuid": "5fb922eb-04d1-4d4c-8bbe-db0f5917744f", 00:09:16.943 "is_configured": true, 00:09:16.943 "data_offset": 0, 00:09:16.943 "data_size": 65536 00:09:16.943 }, 00:09:16.943 { 00:09:16.943 "name": "BaseBdev2", 00:09:16.943 "uuid": "be996da4-bbeb-4920-ac63-e5cc4d61c8d3", 00:09:16.943 "is_configured": true, 00:09:16.943 "data_offset": 0, 00:09:16.943 "data_size": 65536 00:09:16.943 }, 00:09:16.943 { 00:09:16.943 "name": "BaseBdev3", 00:09:16.943 "uuid": "e30d3b71-e051-48ce-b24a-fe1a4b7e76ab", 00:09:16.943 "is_configured": true, 00:09:16.943 "data_offset": 0, 00:09:16.943 "data_size": 65536 00:09:16.943 }, 00:09:16.943 { 00:09:16.943 "name": "BaseBdev4", 00:09:16.943 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.943 "is_configured": false, 00:09:16.943 "data_offset": 0, 00:09:16.943 "data_size": 0 00:09:16.943 } 00:09:16.943 ] 00:09:16.943 }' 00:09:16.943 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:16.943 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.201 [2024-12-07 05:36:50.477194] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:17.201 [2024-12-07 05:36:50.477241] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:17.201 [2024-12-07 05:36:50.477250] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:17.201 [2024-12-07 05:36:50.477534] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:17.201 [2024-12-07 05:36:50.477690] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:17.201 [2024-12-07 05:36:50.477708] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:17.201 [2024-12-07 05:36:50.477902] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:17.201 BaseBdev4 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.201 [ 00:09:17.201 { 00:09:17.201 "name": "BaseBdev4", 00:09:17.201 "aliases": [ 00:09:17.201 "2d8fdcf0-e605-4be1-8493-24cec60ae16b" 00:09:17.201 ], 00:09:17.201 "product_name": "Malloc disk", 00:09:17.201 "block_size": 512, 00:09:17.201 "num_blocks": 65536, 00:09:17.201 "uuid": "2d8fdcf0-e605-4be1-8493-24cec60ae16b", 00:09:17.201 "assigned_rate_limits": { 00:09:17.201 "rw_ios_per_sec": 0, 00:09:17.201 "rw_mbytes_per_sec": 0, 00:09:17.201 "r_mbytes_per_sec": 0, 00:09:17.201 "w_mbytes_per_sec": 0 00:09:17.201 }, 00:09:17.201 "claimed": true, 00:09:17.201 "claim_type": "exclusive_write", 00:09:17.201 "zoned": false, 00:09:17.201 "supported_io_types": { 00:09:17.201 "read": true, 00:09:17.201 "write": true, 00:09:17.201 "unmap": true, 00:09:17.201 "flush": true, 00:09:17.201 "reset": true, 00:09:17.201 "nvme_admin": false, 00:09:17.201 "nvme_io": false, 00:09:17.201 "nvme_io_md": false, 00:09:17.201 "write_zeroes": true, 00:09:17.201 "zcopy": true, 00:09:17.201 "get_zone_info": false, 00:09:17.201 "zone_management": false, 00:09:17.201 "zone_append": false, 00:09:17.201 "compare": false, 00:09:17.201 "compare_and_write": false, 00:09:17.201 "abort": true, 00:09:17.201 "seek_hole": false, 00:09:17.201 "seek_data": false, 00:09:17.201 "copy": true, 00:09:17.201 "nvme_iov_md": false 00:09:17.201 }, 00:09:17.201 "memory_domains": [ 00:09:17.201 { 00:09:17.201 "dma_device_id": "system", 00:09:17.201 "dma_device_type": 1 00:09:17.201 }, 00:09:17.201 { 00:09:17.201 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.201 "dma_device_type": 2 00:09:17.201 } 00:09:17.201 ], 00:09:17.201 "driver_specific": {} 00:09:17.201 } 00:09:17.201 ] 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:17.201 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.459 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:17.459 "name": "Existed_Raid", 00:09:17.459 "uuid": "d4c1f53a-7841-4c67-bbbb-96d2c29bf2f0", 00:09:17.459 "strip_size_kb": 64, 00:09:17.459 "state": "online", 00:09:17.459 "raid_level": "raid0", 00:09:17.459 "superblock": false, 00:09:17.459 "num_base_bdevs": 4, 00:09:17.459 "num_base_bdevs_discovered": 4, 00:09:17.459 "num_base_bdevs_operational": 4, 00:09:17.459 "base_bdevs_list": [ 00:09:17.459 { 00:09:17.459 "name": "BaseBdev1", 00:09:17.459 "uuid": "5fb922eb-04d1-4d4c-8bbe-db0f5917744f", 00:09:17.459 "is_configured": true, 00:09:17.459 "data_offset": 0, 00:09:17.459 "data_size": 65536 00:09:17.459 }, 00:09:17.459 { 00:09:17.459 "name": "BaseBdev2", 00:09:17.459 "uuid": "be996da4-bbeb-4920-ac63-e5cc4d61c8d3", 00:09:17.459 "is_configured": true, 00:09:17.459 "data_offset": 0, 00:09:17.459 "data_size": 65536 00:09:17.459 }, 00:09:17.459 { 00:09:17.459 "name": "BaseBdev3", 00:09:17.459 "uuid": "e30d3b71-e051-48ce-b24a-fe1a4b7e76ab", 00:09:17.459 "is_configured": true, 00:09:17.459 "data_offset": 0, 00:09:17.459 "data_size": 65536 00:09:17.459 }, 00:09:17.459 { 00:09:17.459 "name": "BaseBdev4", 00:09:17.459 "uuid": "2d8fdcf0-e605-4be1-8493-24cec60ae16b", 00:09:17.459 "is_configured": true, 00:09:17.459 "data_offset": 0, 00:09:17.459 "data_size": 65536 00:09:17.459 } 00:09:17.459 ] 00:09:17.459 }' 00:09:17.459 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:17.459 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.719 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:17.719 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:17.719 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:17.719 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:17.719 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:17.719 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:17.719 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:17.719 05:36:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:17.719 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.719 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.719 [2024-12-07 05:36:50.968780] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:17.719 05:36:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.719 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:17.719 "name": "Existed_Raid", 00:09:17.719 "aliases": [ 00:09:17.719 "d4c1f53a-7841-4c67-bbbb-96d2c29bf2f0" 00:09:17.719 ], 00:09:17.719 "product_name": "Raid Volume", 00:09:17.719 "block_size": 512, 00:09:17.719 "num_blocks": 262144, 00:09:17.719 "uuid": "d4c1f53a-7841-4c67-bbbb-96d2c29bf2f0", 00:09:17.719 "assigned_rate_limits": { 00:09:17.719 "rw_ios_per_sec": 0, 00:09:17.719 "rw_mbytes_per_sec": 0, 00:09:17.719 "r_mbytes_per_sec": 0, 00:09:17.719 "w_mbytes_per_sec": 0 00:09:17.719 }, 00:09:17.719 "claimed": false, 00:09:17.719 "zoned": false, 00:09:17.719 "supported_io_types": { 00:09:17.719 "read": true, 00:09:17.719 "write": true, 00:09:17.719 "unmap": true, 00:09:17.719 "flush": true, 00:09:17.719 "reset": true, 00:09:17.719 "nvme_admin": false, 00:09:17.719 "nvme_io": false, 00:09:17.719 "nvme_io_md": false, 00:09:17.719 "write_zeroes": true, 00:09:17.719 "zcopy": false, 00:09:17.719 "get_zone_info": false, 00:09:17.719 "zone_management": false, 00:09:17.719 "zone_append": false, 00:09:17.719 "compare": false, 00:09:17.719 "compare_and_write": false, 00:09:17.719 "abort": false, 00:09:17.719 "seek_hole": false, 00:09:17.719 "seek_data": false, 00:09:17.719 "copy": false, 00:09:17.719 "nvme_iov_md": false 00:09:17.719 }, 00:09:17.719 "memory_domains": [ 00:09:17.719 { 00:09:17.719 "dma_device_id": "system", 00:09:17.719 "dma_device_type": 1 00:09:17.719 }, 00:09:17.719 { 00:09:17.719 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.719 "dma_device_type": 2 00:09:17.719 }, 00:09:17.719 { 00:09:17.719 "dma_device_id": "system", 00:09:17.719 "dma_device_type": 1 00:09:17.719 }, 00:09:17.719 { 00:09:17.719 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.719 "dma_device_type": 2 00:09:17.719 }, 00:09:17.719 { 00:09:17.719 "dma_device_id": "system", 00:09:17.719 "dma_device_type": 1 00:09:17.719 }, 00:09:17.719 { 00:09:17.719 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.719 "dma_device_type": 2 00:09:17.719 }, 00:09:17.719 { 00:09:17.719 "dma_device_id": "system", 00:09:17.719 "dma_device_type": 1 00:09:17.719 }, 00:09:17.719 { 00:09:17.719 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.719 "dma_device_type": 2 00:09:17.719 } 00:09:17.719 ], 00:09:17.719 "driver_specific": { 00:09:17.719 "raid": { 00:09:17.719 "uuid": "d4c1f53a-7841-4c67-bbbb-96d2c29bf2f0", 00:09:17.719 "strip_size_kb": 64, 00:09:17.719 "state": "online", 00:09:17.719 "raid_level": "raid0", 00:09:17.719 "superblock": false, 00:09:17.719 "num_base_bdevs": 4, 00:09:17.719 "num_base_bdevs_discovered": 4, 00:09:17.719 "num_base_bdevs_operational": 4, 00:09:17.719 "base_bdevs_list": [ 00:09:17.719 { 00:09:17.719 "name": "BaseBdev1", 00:09:17.719 "uuid": "5fb922eb-04d1-4d4c-8bbe-db0f5917744f", 00:09:17.719 "is_configured": true, 00:09:17.719 "data_offset": 0, 00:09:17.719 "data_size": 65536 00:09:17.719 }, 00:09:17.719 { 00:09:17.719 "name": "BaseBdev2", 00:09:17.719 "uuid": "be996da4-bbeb-4920-ac63-e5cc4d61c8d3", 00:09:17.719 "is_configured": true, 00:09:17.719 "data_offset": 0, 00:09:17.719 "data_size": 65536 00:09:17.719 }, 00:09:17.719 { 00:09:17.719 "name": "BaseBdev3", 00:09:17.719 "uuid": "e30d3b71-e051-48ce-b24a-fe1a4b7e76ab", 00:09:17.719 "is_configured": true, 00:09:17.719 "data_offset": 0, 00:09:17.719 "data_size": 65536 00:09:17.719 }, 00:09:17.719 { 00:09:17.719 "name": "BaseBdev4", 00:09:17.719 "uuid": "2d8fdcf0-e605-4be1-8493-24cec60ae16b", 00:09:17.719 "is_configured": true, 00:09:17.719 "data_offset": 0, 00:09:17.719 "data_size": 65536 00:09:17.719 } 00:09:17.719 ] 00:09:17.719 } 00:09:17.719 } 00:09:17.719 }' 00:09:17.719 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:17.719 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:17.719 BaseBdev2 00:09:17.719 BaseBdev3 00:09:17.719 BaseBdev4' 00:09:17.719 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.979 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:17.979 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.980 [2024-12-07 05:36:51.291921] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:17.980 [2024-12-07 05:36:51.291952] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:17.980 [2024-12-07 05:36:51.292008] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.980 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.239 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.239 "name": "Existed_Raid", 00:09:18.239 "uuid": "d4c1f53a-7841-4c67-bbbb-96d2c29bf2f0", 00:09:18.239 "strip_size_kb": 64, 00:09:18.239 "state": "offline", 00:09:18.239 "raid_level": "raid0", 00:09:18.239 "superblock": false, 00:09:18.239 "num_base_bdevs": 4, 00:09:18.239 "num_base_bdevs_discovered": 3, 00:09:18.239 "num_base_bdevs_operational": 3, 00:09:18.239 "base_bdevs_list": [ 00:09:18.239 { 00:09:18.239 "name": null, 00:09:18.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.239 "is_configured": false, 00:09:18.239 "data_offset": 0, 00:09:18.239 "data_size": 65536 00:09:18.239 }, 00:09:18.239 { 00:09:18.239 "name": "BaseBdev2", 00:09:18.239 "uuid": "be996da4-bbeb-4920-ac63-e5cc4d61c8d3", 00:09:18.239 "is_configured": true, 00:09:18.239 "data_offset": 0, 00:09:18.239 "data_size": 65536 00:09:18.239 }, 00:09:18.239 { 00:09:18.240 "name": "BaseBdev3", 00:09:18.240 "uuid": "e30d3b71-e051-48ce-b24a-fe1a4b7e76ab", 00:09:18.240 "is_configured": true, 00:09:18.240 "data_offset": 0, 00:09:18.240 "data_size": 65536 00:09:18.240 }, 00:09:18.240 { 00:09:18.240 "name": "BaseBdev4", 00:09:18.240 "uuid": "2d8fdcf0-e605-4be1-8493-24cec60ae16b", 00:09:18.240 "is_configured": true, 00:09:18.240 "data_offset": 0, 00:09:18.240 "data_size": 65536 00:09:18.240 } 00:09:18.240 ] 00:09:18.240 }' 00:09:18.240 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.240 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.499 [2024-12-07 05:36:51.766538] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.499 [2024-12-07 05:36:51.841796] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.499 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.759 [2024-12-07 05:36:51.901016] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:18.759 [2024-12-07 05:36:51.901114] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.759 BaseBdev2 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.759 05:36:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.759 [ 00:09:18.759 { 00:09:18.759 "name": "BaseBdev2", 00:09:18.759 "aliases": [ 00:09:18.759 "f9f34f4b-7386-4a78-adf0-cf6068fa852d" 00:09:18.759 ], 00:09:18.759 "product_name": "Malloc disk", 00:09:18.759 "block_size": 512, 00:09:18.759 "num_blocks": 65536, 00:09:18.759 "uuid": "f9f34f4b-7386-4a78-adf0-cf6068fa852d", 00:09:18.759 "assigned_rate_limits": { 00:09:18.759 "rw_ios_per_sec": 0, 00:09:18.759 "rw_mbytes_per_sec": 0, 00:09:18.759 "r_mbytes_per_sec": 0, 00:09:18.759 "w_mbytes_per_sec": 0 00:09:18.759 }, 00:09:18.759 "claimed": false, 00:09:18.759 "zoned": false, 00:09:18.759 "supported_io_types": { 00:09:18.759 "read": true, 00:09:18.759 "write": true, 00:09:18.759 "unmap": true, 00:09:18.759 "flush": true, 00:09:18.759 "reset": true, 00:09:18.759 "nvme_admin": false, 00:09:18.759 "nvme_io": false, 00:09:18.759 "nvme_io_md": false, 00:09:18.759 "write_zeroes": true, 00:09:18.759 "zcopy": true, 00:09:18.759 "get_zone_info": false, 00:09:18.759 "zone_management": false, 00:09:18.759 "zone_append": false, 00:09:18.759 "compare": false, 00:09:18.759 "compare_and_write": false, 00:09:18.759 "abort": true, 00:09:18.759 "seek_hole": false, 00:09:18.759 "seek_data": false, 00:09:18.759 "copy": true, 00:09:18.759 "nvme_iov_md": false 00:09:18.759 }, 00:09:18.759 "memory_domains": [ 00:09:18.759 { 00:09:18.759 "dma_device_id": "system", 00:09:18.759 "dma_device_type": 1 00:09:18.759 }, 00:09:18.759 { 00:09:18.759 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.759 "dma_device_type": 2 00:09:18.759 } 00:09:18.759 ], 00:09:18.759 "driver_specific": {} 00:09:18.759 } 00:09:18.759 ] 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.759 BaseBdev3 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.759 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.759 [ 00:09:18.759 { 00:09:18.759 "name": "BaseBdev3", 00:09:18.759 "aliases": [ 00:09:18.759 "2dca007b-9965-4ce1-8887-23e149e7f965" 00:09:18.759 ], 00:09:18.760 "product_name": "Malloc disk", 00:09:18.760 "block_size": 512, 00:09:18.760 "num_blocks": 65536, 00:09:18.760 "uuid": "2dca007b-9965-4ce1-8887-23e149e7f965", 00:09:18.760 "assigned_rate_limits": { 00:09:18.760 "rw_ios_per_sec": 0, 00:09:18.760 "rw_mbytes_per_sec": 0, 00:09:18.760 "r_mbytes_per_sec": 0, 00:09:18.760 "w_mbytes_per_sec": 0 00:09:18.760 }, 00:09:18.760 "claimed": false, 00:09:18.760 "zoned": false, 00:09:18.760 "supported_io_types": { 00:09:18.760 "read": true, 00:09:18.760 "write": true, 00:09:18.760 "unmap": true, 00:09:18.760 "flush": true, 00:09:18.760 "reset": true, 00:09:18.760 "nvme_admin": false, 00:09:18.760 "nvme_io": false, 00:09:18.760 "nvme_io_md": false, 00:09:18.760 "write_zeroes": true, 00:09:18.760 "zcopy": true, 00:09:18.760 "get_zone_info": false, 00:09:18.760 "zone_management": false, 00:09:18.760 "zone_append": false, 00:09:18.760 "compare": false, 00:09:18.760 "compare_and_write": false, 00:09:18.760 "abort": true, 00:09:18.760 "seek_hole": false, 00:09:18.760 "seek_data": false, 00:09:18.760 "copy": true, 00:09:18.760 "nvme_iov_md": false 00:09:18.760 }, 00:09:18.760 "memory_domains": [ 00:09:18.760 { 00:09:18.760 "dma_device_id": "system", 00:09:18.760 "dma_device_type": 1 00:09:18.760 }, 00:09:18.760 { 00:09:18.760 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.760 "dma_device_type": 2 00:09:18.760 } 00:09:18.760 ], 00:09:18.760 "driver_specific": {} 00:09:18.760 } 00:09:18.760 ] 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.760 BaseBdev4 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.760 [ 00:09:18.760 { 00:09:18.760 "name": "BaseBdev4", 00:09:18.760 "aliases": [ 00:09:18.760 "f0666091-6eef-4ef6-b7a7-11f02f517d01" 00:09:18.760 ], 00:09:18.760 "product_name": "Malloc disk", 00:09:18.760 "block_size": 512, 00:09:18.760 "num_blocks": 65536, 00:09:18.760 "uuid": "f0666091-6eef-4ef6-b7a7-11f02f517d01", 00:09:18.760 "assigned_rate_limits": { 00:09:18.760 "rw_ios_per_sec": 0, 00:09:18.760 "rw_mbytes_per_sec": 0, 00:09:18.760 "r_mbytes_per_sec": 0, 00:09:18.760 "w_mbytes_per_sec": 0 00:09:18.760 }, 00:09:18.760 "claimed": false, 00:09:18.760 "zoned": false, 00:09:18.760 "supported_io_types": { 00:09:18.760 "read": true, 00:09:18.760 "write": true, 00:09:18.760 "unmap": true, 00:09:18.760 "flush": true, 00:09:18.760 "reset": true, 00:09:18.760 "nvme_admin": false, 00:09:18.760 "nvme_io": false, 00:09:18.760 "nvme_io_md": false, 00:09:18.760 "write_zeroes": true, 00:09:18.760 "zcopy": true, 00:09:18.760 "get_zone_info": false, 00:09:18.760 "zone_management": false, 00:09:18.760 "zone_append": false, 00:09:18.760 "compare": false, 00:09:18.760 "compare_and_write": false, 00:09:18.760 "abort": true, 00:09:18.760 "seek_hole": false, 00:09:18.760 "seek_data": false, 00:09:18.760 "copy": true, 00:09:18.760 "nvme_iov_md": false 00:09:18.760 }, 00:09:18.760 "memory_domains": [ 00:09:18.760 { 00:09:18.760 "dma_device_id": "system", 00:09:18.760 "dma_device_type": 1 00:09:18.760 }, 00:09:18.760 { 00:09:18.760 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.760 "dma_device_type": 2 00:09:18.760 } 00:09:18.760 ], 00:09:18.760 "driver_specific": {} 00:09:18.760 } 00:09:18.760 ] 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:18.760 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.020 [2024-12-07 05:36:52.129987] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:19.020 [2024-12-07 05:36:52.130075] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:19.020 [2024-12-07 05:36:52.130139] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:19.020 [2024-12-07 05:36:52.132033] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:19.020 [2024-12-07 05:36:52.132124] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.020 "name": "Existed_Raid", 00:09:19.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.020 "strip_size_kb": 64, 00:09:19.020 "state": "configuring", 00:09:19.020 "raid_level": "raid0", 00:09:19.020 "superblock": false, 00:09:19.020 "num_base_bdevs": 4, 00:09:19.020 "num_base_bdevs_discovered": 3, 00:09:19.020 "num_base_bdevs_operational": 4, 00:09:19.020 "base_bdevs_list": [ 00:09:19.020 { 00:09:19.020 "name": "BaseBdev1", 00:09:19.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.020 "is_configured": false, 00:09:19.020 "data_offset": 0, 00:09:19.020 "data_size": 0 00:09:19.020 }, 00:09:19.020 { 00:09:19.020 "name": "BaseBdev2", 00:09:19.020 "uuid": "f9f34f4b-7386-4a78-adf0-cf6068fa852d", 00:09:19.020 "is_configured": true, 00:09:19.020 "data_offset": 0, 00:09:19.020 "data_size": 65536 00:09:19.020 }, 00:09:19.020 { 00:09:19.020 "name": "BaseBdev3", 00:09:19.020 "uuid": "2dca007b-9965-4ce1-8887-23e149e7f965", 00:09:19.020 "is_configured": true, 00:09:19.020 "data_offset": 0, 00:09:19.020 "data_size": 65536 00:09:19.020 }, 00:09:19.020 { 00:09:19.020 "name": "BaseBdev4", 00:09:19.020 "uuid": "f0666091-6eef-4ef6-b7a7-11f02f517d01", 00:09:19.020 "is_configured": true, 00:09:19.020 "data_offset": 0, 00:09:19.020 "data_size": 65536 00:09:19.020 } 00:09:19.020 ] 00:09:19.020 }' 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.020 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.280 [2024-12-07 05:36:52.585232] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.280 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.280 "name": "Existed_Raid", 00:09:19.280 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.280 "strip_size_kb": 64, 00:09:19.280 "state": "configuring", 00:09:19.280 "raid_level": "raid0", 00:09:19.280 "superblock": false, 00:09:19.280 "num_base_bdevs": 4, 00:09:19.280 "num_base_bdevs_discovered": 2, 00:09:19.280 "num_base_bdevs_operational": 4, 00:09:19.280 "base_bdevs_list": [ 00:09:19.280 { 00:09:19.280 "name": "BaseBdev1", 00:09:19.280 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.280 "is_configured": false, 00:09:19.280 "data_offset": 0, 00:09:19.280 "data_size": 0 00:09:19.280 }, 00:09:19.280 { 00:09:19.280 "name": null, 00:09:19.280 "uuid": "f9f34f4b-7386-4a78-adf0-cf6068fa852d", 00:09:19.280 "is_configured": false, 00:09:19.280 "data_offset": 0, 00:09:19.280 "data_size": 65536 00:09:19.281 }, 00:09:19.281 { 00:09:19.281 "name": "BaseBdev3", 00:09:19.281 "uuid": "2dca007b-9965-4ce1-8887-23e149e7f965", 00:09:19.281 "is_configured": true, 00:09:19.281 "data_offset": 0, 00:09:19.281 "data_size": 65536 00:09:19.281 }, 00:09:19.281 { 00:09:19.281 "name": "BaseBdev4", 00:09:19.281 "uuid": "f0666091-6eef-4ef6-b7a7-11f02f517d01", 00:09:19.281 "is_configured": true, 00:09:19.281 "data_offset": 0, 00:09:19.281 "data_size": 65536 00:09:19.281 } 00:09:19.281 ] 00:09:19.281 }' 00:09:19.281 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.281 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.847 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.847 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.847 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.847 05:36:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:19.847 05:36:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.847 [2024-12-07 05:36:53.039458] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:19.847 BaseBdev1 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.847 [ 00:09:19.847 { 00:09:19.847 "name": "BaseBdev1", 00:09:19.847 "aliases": [ 00:09:19.847 "1fac95a7-4a8c-4f55-84b9-b3366aa518fd" 00:09:19.847 ], 00:09:19.847 "product_name": "Malloc disk", 00:09:19.847 "block_size": 512, 00:09:19.847 "num_blocks": 65536, 00:09:19.847 "uuid": "1fac95a7-4a8c-4f55-84b9-b3366aa518fd", 00:09:19.847 "assigned_rate_limits": { 00:09:19.847 "rw_ios_per_sec": 0, 00:09:19.847 "rw_mbytes_per_sec": 0, 00:09:19.847 "r_mbytes_per_sec": 0, 00:09:19.847 "w_mbytes_per_sec": 0 00:09:19.847 }, 00:09:19.847 "claimed": true, 00:09:19.847 "claim_type": "exclusive_write", 00:09:19.847 "zoned": false, 00:09:19.847 "supported_io_types": { 00:09:19.847 "read": true, 00:09:19.847 "write": true, 00:09:19.847 "unmap": true, 00:09:19.847 "flush": true, 00:09:19.847 "reset": true, 00:09:19.847 "nvme_admin": false, 00:09:19.847 "nvme_io": false, 00:09:19.847 "nvme_io_md": false, 00:09:19.847 "write_zeroes": true, 00:09:19.847 "zcopy": true, 00:09:19.847 "get_zone_info": false, 00:09:19.847 "zone_management": false, 00:09:19.847 "zone_append": false, 00:09:19.847 "compare": false, 00:09:19.847 "compare_and_write": false, 00:09:19.847 "abort": true, 00:09:19.847 "seek_hole": false, 00:09:19.847 "seek_data": false, 00:09:19.847 "copy": true, 00:09:19.847 "nvme_iov_md": false 00:09:19.847 }, 00:09:19.847 "memory_domains": [ 00:09:19.847 { 00:09:19.847 "dma_device_id": "system", 00:09:19.847 "dma_device_type": 1 00:09:19.847 }, 00:09:19.847 { 00:09:19.847 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.847 "dma_device_type": 2 00:09:19.847 } 00:09:19.847 ], 00:09:19.847 "driver_specific": {} 00:09:19.847 } 00:09:19.847 ] 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.847 "name": "Existed_Raid", 00:09:19.847 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.847 "strip_size_kb": 64, 00:09:19.847 "state": "configuring", 00:09:19.847 "raid_level": "raid0", 00:09:19.847 "superblock": false, 00:09:19.847 "num_base_bdevs": 4, 00:09:19.847 "num_base_bdevs_discovered": 3, 00:09:19.847 "num_base_bdevs_operational": 4, 00:09:19.847 "base_bdevs_list": [ 00:09:19.847 { 00:09:19.847 "name": "BaseBdev1", 00:09:19.847 "uuid": "1fac95a7-4a8c-4f55-84b9-b3366aa518fd", 00:09:19.847 "is_configured": true, 00:09:19.847 "data_offset": 0, 00:09:19.847 "data_size": 65536 00:09:19.847 }, 00:09:19.847 { 00:09:19.847 "name": null, 00:09:19.847 "uuid": "f9f34f4b-7386-4a78-adf0-cf6068fa852d", 00:09:19.847 "is_configured": false, 00:09:19.847 "data_offset": 0, 00:09:19.847 "data_size": 65536 00:09:19.847 }, 00:09:19.847 { 00:09:19.847 "name": "BaseBdev3", 00:09:19.847 "uuid": "2dca007b-9965-4ce1-8887-23e149e7f965", 00:09:19.847 "is_configured": true, 00:09:19.847 "data_offset": 0, 00:09:19.847 "data_size": 65536 00:09:19.847 }, 00:09:19.847 { 00:09:19.847 "name": "BaseBdev4", 00:09:19.847 "uuid": "f0666091-6eef-4ef6-b7a7-11f02f517d01", 00:09:19.847 "is_configured": true, 00:09:19.847 "data_offset": 0, 00:09:19.847 "data_size": 65536 00:09:19.847 } 00:09:19.847 ] 00:09:19.847 }' 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.847 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.414 [2024-12-07 05:36:53.534663] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.414 "name": "Existed_Raid", 00:09:20.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.414 "strip_size_kb": 64, 00:09:20.414 "state": "configuring", 00:09:20.414 "raid_level": "raid0", 00:09:20.414 "superblock": false, 00:09:20.414 "num_base_bdevs": 4, 00:09:20.414 "num_base_bdevs_discovered": 2, 00:09:20.414 "num_base_bdevs_operational": 4, 00:09:20.414 "base_bdevs_list": [ 00:09:20.414 { 00:09:20.414 "name": "BaseBdev1", 00:09:20.414 "uuid": "1fac95a7-4a8c-4f55-84b9-b3366aa518fd", 00:09:20.414 "is_configured": true, 00:09:20.414 "data_offset": 0, 00:09:20.414 "data_size": 65536 00:09:20.414 }, 00:09:20.414 { 00:09:20.414 "name": null, 00:09:20.414 "uuid": "f9f34f4b-7386-4a78-adf0-cf6068fa852d", 00:09:20.414 "is_configured": false, 00:09:20.414 "data_offset": 0, 00:09:20.414 "data_size": 65536 00:09:20.414 }, 00:09:20.414 { 00:09:20.414 "name": null, 00:09:20.414 "uuid": "2dca007b-9965-4ce1-8887-23e149e7f965", 00:09:20.414 "is_configured": false, 00:09:20.414 "data_offset": 0, 00:09:20.414 "data_size": 65536 00:09:20.414 }, 00:09:20.414 { 00:09:20.414 "name": "BaseBdev4", 00:09:20.414 "uuid": "f0666091-6eef-4ef6-b7a7-11f02f517d01", 00:09:20.414 "is_configured": true, 00:09:20.414 "data_offset": 0, 00:09:20.414 "data_size": 65536 00:09:20.414 } 00:09:20.414 ] 00:09:20.414 }' 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.414 05:36:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.672 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.672 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:20.672 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.672 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.672 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.929 [2024-12-07 05:36:54.057820] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.929 "name": "Existed_Raid", 00:09:20.929 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.929 "strip_size_kb": 64, 00:09:20.929 "state": "configuring", 00:09:20.929 "raid_level": "raid0", 00:09:20.929 "superblock": false, 00:09:20.929 "num_base_bdevs": 4, 00:09:20.929 "num_base_bdevs_discovered": 3, 00:09:20.929 "num_base_bdevs_operational": 4, 00:09:20.929 "base_bdevs_list": [ 00:09:20.929 { 00:09:20.929 "name": "BaseBdev1", 00:09:20.929 "uuid": "1fac95a7-4a8c-4f55-84b9-b3366aa518fd", 00:09:20.929 "is_configured": true, 00:09:20.929 "data_offset": 0, 00:09:20.929 "data_size": 65536 00:09:20.929 }, 00:09:20.929 { 00:09:20.929 "name": null, 00:09:20.929 "uuid": "f9f34f4b-7386-4a78-adf0-cf6068fa852d", 00:09:20.929 "is_configured": false, 00:09:20.929 "data_offset": 0, 00:09:20.929 "data_size": 65536 00:09:20.929 }, 00:09:20.929 { 00:09:20.929 "name": "BaseBdev3", 00:09:20.929 "uuid": "2dca007b-9965-4ce1-8887-23e149e7f965", 00:09:20.929 "is_configured": true, 00:09:20.929 "data_offset": 0, 00:09:20.929 "data_size": 65536 00:09:20.929 }, 00:09:20.929 { 00:09:20.929 "name": "BaseBdev4", 00:09:20.929 "uuid": "f0666091-6eef-4ef6-b7a7-11f02f517d01", 00:09:20.929 "is_configured": true, 00:09:20.929 "data_offset": 0, 00:09:20.929 "data_size": 65536 00:09:20.929 } 00:09:20.929 ] 00:09:20.929 }' 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.929 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.188 [2024-12-07 05:36:54.537028] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.188 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.447 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:21.447 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.447 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.447 "name": "Existed_Raid", 00:09:21.447 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.447 "strip_size_kb": 64, 00:09:21.447 "state": "configuring", 00:09:21.447 "raid_level": "raid0", 00:09:21.447 "superblock": false, 00:09:21.447 "num_base_bdevs": 4, 00:09:21.447 "num_base_bdevs_discovered": 2, 00:09:21.447 "num_base_bdevs_operational": 4, 00:09:21.447 "base_bdevs_list": [ 00:09:21.447 { 00:09:21.447 "name": null, 00:09:21.447 "uuid": "1fac95a7-4a8c-4f55-84b9-b3366aa518fd", 00:09:21.447 "is_configured": false, 00:09:21.447 "data_offset": 0, 00:09:21.447 "data_size": 65536 00:09:21.447 }, 00:09:21.447 { 00:09:21.447 "name": null, 00:09:21.447 "uuid": "f9f34f4b-7386-4a78-adf0-cf6068fa852d", 00:09:21.447 "is_configured": false, 00:09:21.447 "data_offset": 0, 00:09:21.447 "data_size": 65536 00:09:21.447 }, 00:09:21.447 { 00:09:21.447 "name": "BaseBdev3", 00:09:21.447 "uuid": "2dca007b-9965-4ce1-8887-23e149e7f965", 00:09:21.447 "is_configured": true, 00:09:21.447 "data_offset": 0, 00:09:21.447 "data_size": 65536 00:09:21.447 }, 00:09:21.447 { 00:09:21.447 "name": "BaseBdev4", 00:09:21.447 "uuid": "f0666091-6eef-4ef6-b7a7-11f02f517d01", 00:09:21.447 "is_configured": true, 00:09:21.447 "data_offset": 0, 00:09:21.447 "data_size": 65536 00:09:21.447 } 00:09:21.447 ] 00:09:21.447 }' 00:09:21.447 05:36:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.447 05:36:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.704 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.705 [2024-12-07 05:36:55.058888] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.705 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:21.962 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.962 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.962 "name": "Existed_Raid", 00:09:21.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.962 "strip_size_kb": 64, 00:09:21.962 "state": "configuring", 00:09:21.962 "raid_level": "raid0", 00:09:21.962 "superblock": false, 00:09:21.962 "num_base_bdevs": 4, 00:09:21.962 "num_base_bdevs_discovered": 3, 00:09:21.962 "num_base_bdevs_operational": 4, 00:09:21.962 "base_bdevs_list": [ 00:09:21.962 { 00:09:21.962 "name": null, 00:09:21.962 "uuid": "1fac95a7-4a8c-4f55-84b9-b3366aa518fd", 00:09:21.962 "is_configured": false, 00:09:21.962 "data_offset": 0, 00:09:21.962 "data_size": 65536 00:09:21.962 }, 00:09:21.963 { 00:09:21.963 "name": "BaseBdev2", 00:09:21.963 "uuid": "f9f34f4b-7386-4a78-adf0-cf6068fa852d", 00:09:21.963 "is_configured": true, 00:09:21.963 "data_offset": 0, 00:09:21.963 "data_size": 65536 00:09:21.963 }, 00:09:21.963 { 00:09:21.963 "name": "BaseBdev3", 00:09:21.963 "uuid": "2dca007b-9965-4ce1-8887-23e149e7f965", 00:09:21.963 "is_configured": true, 00:09:21.963 "data_offset": 0, 00:09:21.963 "data_size": 65536 00:09:21.963 }, 00:09:21.963 { 00:09:21.963 "name": "BaseBdev4", 00:09:21.963 "uuid": "f0666091-6eef-4ef6-b7a7-11f02f517d01", 00:09:21.963 "is_configured": true, 00:09:21.963 "data_offset": 0, 00:09:21.963 "data_size": 65536 00:09:21.963 } 00:09:21.963 ] 00:09:21.963 }' 00:09:21.963 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.963 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.242 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 1fac95a7-4a8c-4f55-84b9-b3366aa518fd 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.243 [2024-12-07 05:36:55.581029] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:22.243 [2024-12-07 05:36:55.581133] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:22.243 [2024-12-07 05:36:55.581159] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:22.243 [2024-12-07 05:36:55.581483] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:22.243 [2024-12-07 05:36:55.581646] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:22.243 [2024-12-07 05:36:55.581688] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:22.243 [2024-12-07 05:36:55.581895] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:22.243 NewBaseBdev 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.243 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.501 [ 00:09:22.501 { 00:09:22.501 "name": "NewBaseBdev", 00:09:22.501 "aliases": [ 00:09:22.501 "1fac95a7-4a8c-4f55-84b9-b3366aa518fd" 00:09:22.501 ], 00:09:22.501 "product_name": "Malloc disk", 00:09:22.501 "block_size": 512, 00:09:22.501 "num_blocks": 65536, 00:09:22.501 "uuid": "1fac95a7-4a8c-4f55-84b9-b3366aa518fd", 00:09:22.501 "assigned_rate_limits": { 00:09:22.501 "rw_ios_per_sec": 0, 00:09:22.501 "rw_mbytes_per_sec": 0, 00:09:22.501 "r_mbytes_per_sec": 0, 00:09:22.501 "w_mbytes_per_sec": 0 00:09:22.501 }, 00:09:22.501 "claimed": true, 00:09:22.501 "claim_type": "exclusive_write", 00:09:22.501 "zoned": false, 00:09:22.501 "supported_io_types": { 00:09:22.501 "read": true, 00:09:22.501 "write": true, 00:09:22.501 "unmap": true, 00:09:22.501 "flush": true, 00:09:22.501 "reset": true, 00:09:22.501 "nvme_admin": false, 00:09:22.501 "nvme_io": false, 00:09:22.501 "nvme_io_md": false, 00:09:22.501 "write_zeroes": true, 00:09:22.501 "zcopy": true, 00:09:22.501 "get_zone_info": false, 00:09:22.501 "zone_management": false, 00:09:22.501 "zone_append": false, 00:09:22.501 "compare": false, 00:09:22.501 "compare_and_write": false, 00:09:22.501 "abort": true, 00:09:22.501 "seek_hole": false, 00:09:22.501 "seek_data": false, 00:09:22.501 "copy": true, 00:09:22.501 "nvme_iov_md": false 00:09:22.501 }, 00:09:22.501 "memory_domains": [ 00:09:22.501 { 00:09:22.501 "dma_device_id": "system", 00:09:22.501 "dma_device_type": 1 00:09:22.501 }, 00:09:22.502 { 00:09:22.502 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.502 "dma_device_type": 2 00:09:22.502 } 00:09:22.502 ], 00:09:22.502 "driver_specific": {} 00:09:22.502 } 00:09:22.502 ] 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.502 "name": "Existed_Raid", 00:09:22.502 "uuid": "4601c490-9f41-4cf9-8182-133a4a0bf52b", 00:09:22.502 "strip_size_kb": 64, 00:09:22.502 "state": "online", 00:09:22.502 "raid_level": "raid0", 00:09:22.502 "superblock": false, 00:09:22.502 "num_base_bdevs": 4, 00:09:22.502 "num_base_bdevs_discovered": 4, 00:09:22.502 "num_base_bdevs_operational": 4, 00:09:22.502 "base_bdevs_list": [ 00:09:22.502 { 00:09:22.502 "name": "NewBaseBdev", 00:09:22.502 "uuid": "1fac95a7-4a8c-4f55-84b9-b3366aa518fd", 00:09:22.502 "is_configured": true, 00:09:22.502 "data_offset": 0, 00:09:22.502 "data_size": 65536 00:09:22.502 }, 00:09:22.502 { 00:09:22.502 "name": "BaseBdev2", 00:09:22.502 "uuid": "f9f34f4b-7386-4a78-adf0-cf6068fa852d", 00:09:22.502 "is_configured": true, 00:09:22.502 "data_offset": 0, 00:09:22.502 "data_size": 65536 00:09:22.502 }, 00:09:22.502 { 00:09:22.502 "name": "BaseBdev3", 00:09:22.502 "uuid": "2dca007b-9965-4ce1-8887-23e149e7f965", 00:09:22.502 "is_configured": true, 00:09:22.502 "data_offset": 0, 00:09:22.502 "data_size": 65536 00:09:22.502 }, 00:09:22.502 { 00:09:22.502 "name": "BaseBdev4", 00:09:22.502 "uuid": "f0666091-6eef-4ef6-b7a7-11f02f517d01", 00:09:22.502 "is_configured": true, 00:09:22.502 "data_offset": 0, 00:09:22.502 "data_size": 65536 00:09:22.502 } 00:09:22.502 ] 00:09:22.502 }' 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.502 05:36:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.761 [2024-12-07 05:36:56.032660] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:22.761 "name": "Existed_Raid", 00:09:22.761 "aliases": [ 00:09:22.761 "4601c490-9f41-4cf9-8182-133a4a0bf52b" 00:09:22.761 ], 00:09:22.761 "product_name": "Raid Volume", 00:09:22.761 "block_size": 512, 00:09:22.761 "num_blocks": 262144, 00:09:22.761 "uuid": "4601c490-9f41-4cf9-8182-133a4a0bf52b", 00:09:22.761 "assigned_rate_limits": { 00:09:22.761 "rw_ios_per_sec": 0, 00:09:22.761 "rw_mbytes_per_sec": 0, 00:09:22.761 "r_mbytes_per_sec": 0, 00:09:22.761 "w_mbytes_per_sec": 0 00:09:22.761 }, 00:09:22.761 "claimed": false, 00:09:22.761 "zoned": false, 00:09:22.761 "supported_io_types": { 00:09:22.761 "read": true, 00:09:22.761 "write": true, 00:09:22.761 "unmap": true, 00:09:22.761 "flush": true, 00:09:22.761 "reset": true, 00:09:22.761 "nvme_admin": false, 00:09:22.761 "nvme_io": false, 00:09:22.761 "nvme_io_md": false, 00:09:22.761 "write_zeroes": true, 00:09:22.761 "zcopy": false, 00:09:22.761 "get_zone_info": false, 00:09:22.761 "zone_management": false, 00:09:22.761 "zone_append": false, 00:09:22.761 "compare": false, 00:09:22.761 "compare_and_write": false, 00:09:22.761 "abort": false, 00:09:22.761 "seek_hole": false, 00:09:22.761 "seek_data": false, 00:09:22.761 "copy": false, 00:09:22.761 "nvme_iov_md": false 00:09:22.761 }, 00:09:22.761 "memory_domains": [ 00:09:22.761 { 00:09:22.761 "dma_device_id": "system", 00:09:22.761 "dma_device_type": 1 00:09:22.761 }, 00:09:22.761 { 00:09:22.761 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.761 "dma_device_type": 2 00:09:22.761 }, 00:09:22.761 { 00:09:22.761 "dma_device_id": "system", 00:09:22.761 "dma_device_type": 1 00:09:22.761 }, 00:09:22.761 { 00:09:22.761 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.761 "dma_device_type": 2 00:09:22.761 }, 00:09:22.761 { 00:09:22.761 "dma_device_id": "system", 00:09:22.761 "dma_device_type": 1 00:09:22.761 }, 00:09:22.761 { 00:09:22.761 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.761 "dma_device_type": 2 00:09:22.761 }, 00:09:22.761 { 00:09:22.761 "dma_device_id": "system", 00:09:22.761 "dma_device_type": 1 00:09:22.761 }, 00:09:22.761 { 00:09:22.761 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.761 "dma_device_type": 2 00:09:22.761 } 00:09:22.761 ], 00:09:22.761 "driver_specific": { 00:09:22.761 "raid": { 00:09:22.761 "uuid": "4601c490-9f41-4cf9-8182-133a4a0bf52b", 00:09:22.761 "strip_size_kb": 64, 00:09:22.761 "state": "online", 00:09:22.761 "raid_level": "raid0", 00:09:22.761 "superblock": false, 00:09:22.761 "num_base_bdevs": 4, 00:09:22.761 "num_base_bdevs_discovered": 4, 00:09:22.761 "num_base_bdevs_operational": 4, 00:09:22.761 "base_bdevs_list": [ 00:09:22.761 { 00:09:22.761 "name": "NewBaseBdev", 00:09:22.761 "uuid": "1fac95a7-4a8c-4f55-84b9-b3366aa518fd", 00:09:22.761 "is_configured": true, 00:09:22.761 "data_offset": 0, 00:09:22.761 "data_size": 65536 00:09:22.761 }, 00:09:22.761 { 00:09:22.761 "name": "BaseBdev2", 00:09:22.761 "uuid": "f9f34f4b-7386-4a78-adf0-cf6068fa852d", 00:09:22.761 "is_configured": true, 00:09:22.761 "data_offset": 0, 00:09:22.761 "data_size": 65536 00:09:22.761 }, 00:09:22.761 { 00:09:22.761 "name": "BaseBdev3", 00:09:22.761 "uuid": "2dca007b-9965-4ce1-8887-23e149e7f965", 00:09:22.761 "is_configured": true, 00:09:22.761 "data_offset": 0, 00:09:22.761 "data_size": 65536 00:09:22.761 }, 00:09:22.761 { 00:09:22.761 "name": "BaseBdev4", 00:09:22.761 "uuid": "f0666091-6eef-4ef6-b7a7-11f02f517d01", 00:09:22.761 "is_configured": true, 00:09:22.761 "data_offset": 0, 00:09:22.761 "data_size": 65536 00:09:22.761 } 00:09:22.761 ] 00:09:22.761 } 00:09:22.761 } 00:09:22.761 }' 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:22.761 BaseBdev2 00:09:22.761 BaseBdev3 00:09:22.761 BaseBdev4' 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.761 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.021 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.022 [2024-12-07 05:36:56.319836] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:23.022 [2024-12-07 05:36:56.319865] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:23.022 [2024-12-07 05:36:56.319941] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:23.022 [2024-12-07 05:36:56.320011] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:23.022 [2024-12-07 05:36:56.320022] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 80008 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 80008 ']' 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 80008 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80008 00:09:23.022 killing process with pid 80008 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80008' 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 80008 00:09:23.022 [2024-12-07 05:36:56.354738] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:23.022 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 80008 00:09:23.281 [2024-12-07 05:36:56.394809] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:23.281 05:36:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:23.281 ************************************ 00:09:23.281 END TEST raid_state_function_test 00:09:23.281 ************************************ 00:09:23.281 00:09:23.281 real 0m9.581s 00:09:23.281 user 0m16.536s 00:09:23.281 sys 0m1.895s 00:09:23.281 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:23.281 05:36:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.541 05:36:56 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:09:23.541 05:36:56 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:23.541 05:36:56 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:23.541 05:36:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:23.541 ************************************ 00:09:23.541 START TEST raid_state_function_test_sb 00:09:23.541 ************************************ 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 true 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=80663 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80663' 00:09:23.541 Process raid pid: 80663 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 80663 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 80663 ']' 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:23.541 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:23.541 05:36:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.541 [2024-12-07 05:36:56.764748] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:09:23.541 [2024-12-07 05:36:56.764954] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:23.801 [2024-12-07 05:36:56.921431] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:23.801 [2024-12-07 05:36:56.947535] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:23.801 [2024-12-07 05:36:56.989706] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:23.801 [2024-12-07 05:36:56.989839] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.390 [2024-12-07 05:36:57.636114] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:24.390 [2024-12-07 05:36:57.636177] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:24.390 [2024-12-07 05:36:57.636203] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:24.390 [2024-12-07 05:36:57.636213] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:24.390 [2024-12-07 05:36:57.636219] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:24.390 [2024-12-07 05:36:57.636232] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:24.390 [2024-12-07 05:36:57.636238] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:24.390 [2024-12-07 05:36:57.636248] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:24.390 "name": "Existed_Raid", 00:09:24.390 "uuid": "f542e6cc-57e7-4ce0-a744-963d2b86cf77", 00:09:24.390 "strip_size_kb": 64, 00:09:24.390 "state": "configuring", 00:09:24.390 "raid_level": "raid0", 00:09:24.390 "superblock": true, 00:09:24.390 "num_base_bdevs": 4, 00:09:24.390 "num_base_bdevs_discovered": 0, 00:09:24.390 "num_base_bdevs_operational": 4, 00:09:24.390 "base_bdevs_list": [ 00:09:24.390 { 00:09:24.390 "name": "BaseBdev1", 00:09:24.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:24.390 "is_configured": false, 00:09:24.390 "data_offset": 0, 00:09:24.390 "data_size": 0 00:09:24.390 }, 00:09:24.390 { 00:09:24.390 "name": "BaseBdev2", 00:09:24.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:24.390 "is_configured": false, 00:09:24.390 "data_offset": 0, 00:09:24.390 "data_size": 0 00:09:24.390 }, 00:09:24.390 { 00:09:24.390 "name": "BaseBdev3", 00:09:24.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:24.390 "is_configured": false, 00:09:24.390 "data_offset": 0, 00:09:24.390 "data_size": 0 00:09:24.390 }, 00:09:24.390 { 00:09:24.390 "name": "BaseBdev4", 00:09:24.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:24.390 "is_configured": false, 00:09:24.390 "data_offset": 0, 00:09:24.390 "data_size": 0 00:09:24.390 } 00:09:24.390 ] 00:09:24.390 }' 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:24.390 05:36:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.001 [2024-12-07 05:36:58.067279] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:25.001 [2024-12-07 05:36:58.067367] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.001 [2024-12-07 05:36:58.079285] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:25.001 [2024-12-07 05:36:58.079368] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:25.001 [2024-12-07 05:36:58.079421] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:25.001 [2024-12-07 05:36:58.079449] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:25.001 [2024-12-07 05:36:58.079479] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:25.001 [2024-12-07 05:36:58.079511] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:25.001 [2024-12-07 05:36:58.079540] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:25.001 [2024-12-07 05:36:58.079572] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.001 [2024-12-07 05:36:58.100314] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:25.001 BaseBdev1 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.001 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.002 [ 00:09:25.002 { 00:09:25.002 "name": "BaseBdev1", 00:09:25.002 "aliases": [ 00:09:25.002 "82564374-df7b-477f-9da3-4176f8d34a9c" 00:09:25.002 ], 00:09:25.002 "product_name": "Malloc disk", 00:09:25.002 "block_size": 512, 00:09:25.002 "num_blocks": 65536, 00:09:25.002 "uuid": "82564374-df7b-477f-9da3-4176f8d34a9c", 00:09:25.002 "assigned_rate_limits": { 00:09:25.002 "rw_ios_per_sec": 0, 00:09:25.002 "rw_mbytes_per_sec": 0, 00:09:25.002 "r_mbytes_per_sec": 0, 00:09:25.002 "w_mbytes_per_sec": 0 00:09:25.002 }, 00:09:25.002 "claimed": true, 00:09:25.002 "claim_type": "exclusive_write", 00:09:25.002 "zoned": false, 00:09:25.002 "supported_io_types": { 00:09:25.002 "read": true, 00:09:25.002 "write": true, 00:09:25.002 "unmap": true, 00:09:25.002 "flush": true, 00:09:25.002 "reset": true, 00:09:25.002 "nvme_admin": false, 00:09:25.002 "nvme_io": false, 00:09:25.002 "nvme_io_md": false, 00:09:25.002 "write_zeroes": true, 00:09:25.002 "zcopy": true, 00:09:25.002 "get_zone_info": false, 00:09:25.002 "zone_management": false, 00:09:25.002 "zone_append": false, 00:09:25.002 "compare": false, 00:09:25.002 "compare_and_write": false, 00:09:25.002 "abort": true, 00:09:25.002 "seek_hole": false, 00:09:25.002 "seek_data": false, 00:09:25.002 "copy": true, 00:09:25.002 "nvme_iov_md": false 00:09:25.002 }, 00:09:25.002 "memory_domains": [ 00:09:25.002 { 00:09:25.002 "dma_device_id": "system", 00:09:25.002 "dma_device_type": 1 00:09:25.002 }, 00:09:25.002 { 00:09:25.002 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.002 "dma_device_type": 2 00:09:25.002 } 00:09:25.002 ], 00:09:25.002 "driver_specific": {} 00:09:25.002 } 00:09:25.002 ] 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:25.002 "name": "Existed_Raid", 00:09:25.002 "uuid": "bded5a41-6a0a-478c-8cb7-ade286bc18df", 00:09:25.002 "strip_size_kb": 64, 00:09:25.002 "state": "configuring", 00:09:25.002 "raid_level": "raid0", 00:09:25.002 "superblock": true, 00:09:25.002 "num_base_bdevs": 4, 00:09:25.002 "num_base_bdevs_discovered": 1, 00:09:25.002 "num_base_bdevs_operational": 4, 00:09:25.002 "base_bdevs_list": [ 00:09:25.002 { 00:09:25.002 "name": "BaseBdev1", 00:09:25.002 "uuid": "82564374-df7b-477f-9da3-4176f8d34a9c", 00:09:25.002 "is_configured": true, 00:09:25.002 "data_offset": 2048, 00:09:25.002 "data_size": 63488 00:09:25.002 }, 00:09:25.002 { 00:09:25.002 "name": "BaseBdev2", 00:09:25.002 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.002 "is_configured": false, 00:09:25.002 "data_offset": 0, 00:09:25.002 "data_size": 0 00:09:25.002 }, 00:09:25.002 { 00:09:25.002 "name": "BaseBdev3", 00:09:25.002 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.002 "is_configured": false, 00:09:25.002 "data_offset": 0, 00:09:25.002 "data_size": 0 00:09:25.002 }, 00:09:25.002 { 00:09:25.002 "name": "BaseBdev4", 00:09:25.002 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.002 "is_configured": false, 00:09:25.002 "data_offset": 0, 00:09:25.002 "data_size": 0 00:09:25.002 } 00:09:25.002 ] 00:09:25.002 }' 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:25.002 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.261 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:25.262 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.262 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.262 [2024-12-07 05:36:58.611516] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:25.262 [2024-12-07 05:36:58.611580] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:25.262 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.262 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:25.262 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.262 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.262 [2024-12-07 05:36:58.623544] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:25.262 [2024-12-07 05:36:58.625570] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:25.262 [2024-12-07 05:36:58.625615] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:25.262 [2024-12-07 05:36:58.625637] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:25.262 [2024-12-07 05:36:58.625647] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:25.262 [2024-12-07 05:36:58.625653] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:25.262 [2024-12-07 05:36:58.625661] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.521 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:25.521 "name": "Existed_Raid", 00:09:25.521 "uuid": "e32e24c6-3a2e-47e4-b02c-be3732706575", 00:09:25.521 "strip_size_kb": 64, 00:09:25.521 "state": "configuring", 00:09:25.521 "raid_level": "raid0", 00:09:25.521 "superblock": true, 00:09:25.521 "num_base_bdevs": 4, 00:09:25.521 "num_base_bdevs_discovered": 1, 00:09:25.521 "num_base_bdevs_operational": 4, 00:09:25.521 "base_bdevs_list": [ 00:09:25.521 { 00:09:25.521 "name": "BaseBdev1", 00:09:25.521 "uuid": "82564374-df7b-477f-9da3-4176f8d34a9c", 00:09:25.521 "is_configured": true, 00:09:25.521 "data_offset": 2048, 00:09:25.521 "data_size": 63488 00:09:25.521 }, 00:09:25.521 { 00:09:25.521 "name": "BaseBdev2", 00:09:25.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.521 "is_configured": false, 00:09:25.521 "data_offset": 0, 00:09:25.521 "data_size": 0 00:09:25.521 }, 00:09:25.521 { 00:09:25.521 "name": "BaseBdev3", 00:09:25.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.521 "is_configured": false, 00:09:25.521 "data_offset": 0, 00:09:25.521 "data_size": 0 00:09:25.521 }, 00:09:25.521 { 00:09:25.521 "name": "BaseBdev4", 00:09:25.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.521 "is_configured": false, 00:09:25.521 "data_offset": 0, 00:09:25.521 "data_size": 0 00:09:25.522 } 00:09:25.522 ] 00:09:25.522 }' 00:09:25.522 05:36:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:25.522 05:36:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.781 [2024-12-07 05:36:59.097569] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:25.781 BaseBdev2 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.781 [ 00:09:25.781 { 00:09:25.781 "name": "BaseBdev2", 00:09:25.781 "aliases": [ 00:09:25.781 "e26fd56a-f928-42bd-b17d-905f5ca082d2" 00:09:25.781 ], 00:09:25.781 "product_name": "Malloc disk", 00:09:25.781 "block_size": 512, 00:09:25.781 "num_blocks": 65536, 00:09:25.781 "uuid": "e26fd56a-f928-42bd-b17d-905f5ca082d2", 00:09:25.781 "assigned_rate_limits": { 00:09:25.781 "rw_ios_per_sec": 0, 00:09:25.781 "rw_mbytes_per_sec": 0, 00:09:25.781 "r_mbytes_per_sec": 0, 00:09:25.781 "w_mbytes_per_sec": 0 00:09:25.781 }, 00:09:25.781 "claimed": true, 00:09:25.781 "claim_type": "exclusive_write", 00:09:25.781 "zoned": false, 00:09:25.781 "supported_io_types": { 00:09:25.781 "read": true, 00:09:25.781 "write": true, 00:09:25.781 "unmap": true, 00:09:25.781 "flush": true, 00:09:25.781 "reset": true, 00:09:25.781 "nvme_admin": false, 00:09:25.781 "nvme_io": false, 00:09:25.781 "nvme_io_md": false, 00:09:25.781 "write_zeroes": true, 00:09:25.781 "zcopy": true, 00:09:25.781 "get_zone_info": false, 00:09:25.781 "zone_management": false, 00:09:25.781 "zone_append": false, 00:09:25.781 "compare": false, 00:09:25.781 "compare_and_write": false, 00:09:25.781 "abort": true, 00:09:25.781 "seek_hole": false, 00:09:25.781 "seek_data": false, 00:09:25.781 "copy": true, 00:09:25.781 "nvme_iov_md": false 00:09:25.781 }, 00:09:25.781 "memory_domains": [ 00:09:25.781 { 00:09:25.781 "dma_device_id": "system", 00:09:25.781 "dma_device_type": 1 00:09:25.781 }, 00:09:25.781 { 00:09:25.781 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.781 "dma_device_type": 2 00:09:25.781 } 00:09:25.781 ], 00:09:25.781 "driver_specific": {} 00:09:25.781 } 00:09:25.781 ] 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.781 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.041 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.041 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.041 "name": "Existed_Raid", 00:09:26.041 "uuid": "e32e24c6-3a2e-47e4-b02c-be3732706575", 00:09:26.041 "strip_size_kb": 64, 00:09:26.041 "state": "configuring", 00:09:26.041 "raid_level": "raid0", 00:09:26.041 "superblock": true, 00:09:26.041 "num_base_bdevs": 4, 00:09:26.041 "num_base_bdevs_discovered": 2, 00:09:26.041 "num_base_bdevs_operational": 4, 00:09:26.041 "base_bdevs_list": [ 00:09:26.041 { 00:09:26.041 "name": "BaseBdev1", 00:09:26.041 "uuid": "82564374-df7b-477f-9da3-4176f8d34a9c", 00:09:26.041 "is_configured": true, 00:09:26.041 "data_offset": 2048, 00:09:26.041 "data_size": 63488 00:09:26.041 }, 00:09:26.041 { 00:09:26.041 "name": "BaseBdev2", 00:09:26.041 "uuid": "e26fd56a-f928-42bd-b17d-905f5ca082d2", 00:09:26.041 "is_configured": true, 00:09:26.041 "data_offset": 2048, 00:09:26.041 "data_size": 63488 00:09:26.041 }, 00:09:26.041 { 00:09:26.041 "name": "BaseBdev3", 00:09:26.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.041 "is_configured": false, 00:09:26.041 "data_offset": 0, 00:09:26.041 "data_size": 0 00:09:26.041 }, 00:09:26.041 { 00:09:26.041 "name": "BaseBdev4", 00:09:26.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.041 "is_configured": false, 00:09:26.041 "data_offset": 0, 00:09:26.041 "data_size": 0 00:09:26.041 } 00:09:26.041 ] 00:09:26.041 }' 00:09:26.041 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.041 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.301 [2024-12-07 05:36:59.550724] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:26.301 BaseBdev3 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.301 [ 00:09:26.301 { 00:09:26.301 "name": "BaseBdev3", 00:09:26.301 "aliases": [ 00:09:26.301 "ed92f78c-6775-4eeb-b44f-0a2be315aa16" 00:09:26.301 ], 00:09:26.301 "product_name": "Malloc disk", 00:09:26.301 "block_size": 512, 00:09:26.301 "num_blocks": 65536, 00:09:26.301 "uuid": "ed92f78c-6775-4eeb-b44f-0a2be315aa16", 00:09:26.301 "assigned_rate_limits": { 00:09:26.301 "rw_ios_per_sec": 0, 00:09:26.301 "rw_mbytes_per_sec": 0, 00:09:26.301 "r_mbytes_per_sec": 0, 00:09:26.301 "w_mbytes_per_sec": 0 00:09:26.301 }, 00:09:26.301 "claimed": true, 00:09:26.301 "claim_type": "exclusive_write", 00:09:26.301 "zoned": false, 00:09:26.301 "supported_io_types": { 00:09:26.301 "read": true, 00:09:26.301 "write": true, 00:09:26.301 "unmap": true, 00:09:26.301 "flush": true, 00:09:26.301 "reset": true, 00:09:26.301 "nvme_admin": false, 00:09:26.301 "nvme_io": false, 00:09:26.301 "nvme_io_md": false, 00:09:26.301 "write_zeroes": true, 00:09:26.301 "zcopy": true, 00:09:26.301 "get_zone_info": false, 00:09:26.301 "zone_management": false, 00:09:26.301 "zone_append": false, 00:09:26.301 "compare": false, 00:09:26.301 "compare_and_write": false, 00:09:26.301 "abort": true, 00:09:26.301 "seek_hole": false, 00:09:26.301 "seek_data": false, 00:09:26.301 "copy": true, 00:09:26.301 "nvme_iov_md": false 00:09:26.301 }, 00:09:26.301 "memory_domains": [ 00:09:26.301 { 00:09:26.301 "dma_device_id": "system", 00:09:26.301 "dma_device_type": 1 00:09:26.301 }, 00:09:26.301 { 00:09:26.301 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:26.301 "dma_device_type": 2 00:09:26.301 } 00:09:26.301 ], 00:09:26.301 "driver_specific": {} 00:09:26.301 } 00:09:26.301 ] 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.301 "name": "Existed_Raid", 00:09:26.301 "uuid": "e32e24c6-3a2e-47e4-b02c-be3732706575", 00:09:26.301 "strip_size_kb": 64, 00:09:26.301 "state": "configuring", 00:09:26.301 "raid_level": "raid0", 00:09:26.301 "superblock": true, 00:09:26.301 "num_base_bdevs": 4, 00:09:26.301 "num_base_bdevs_discovered": 3, 00:09:26.301 "num_base_bdevs_operational": 4, 00:09:26.301 "base_bdevs_list": [ 00:09:26.301 { 00:09:26.301 "name": "BaseBdev1", 00:09:26.301 "uuid": "82564374-df7b-477f-9da3-4176f8d34a9c", 00:09:26.301 "is_configured": true, 00:09:26.301 "data_offset": 2048, 00:09:26.301 "data_size": 63488 00:09:26.301 }, 00:09:26.301 { 00:09:26.301 "name": "BaseBdev2", 00:09:26.301 "uuid": "e26fd56a-f928-42bd-b17d-905f5ca082d2", 00:09:26.301 "is_configured": true, 00:09:26.301 "data_offset": 2048, 00:09:26.301 "data_size": 63488 00:09:26.301 }, 00:09:26.301 { 00:09:26.301 "name": "BaseBdev3", 00:09:26.301 "uuid": "ed92f78c-6775-4eeb-b44f-0a2be315aa16", 00:09:26.301 "is_configured": true, 00:09:26.301 "data_offset": 2048, 00:09:26.301 "data_size": 63488 00:09:26.301 }, 00:09:26.301 { 00:09:26.301 "name": "BaseBdev4", 00:09:26.301 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.301 "is_configured": false, 00:09:26.301 "data_offset": 0, 00:09:26.301 "data_size": 0 00:09:26.301 } 00:09:26.301 ] 00:09:26.301 }' 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.301 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.869 05:36:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:26.869 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.869 05:36:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.869 [2024-12-07 05:37:00.005082] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:26.869 [2024-12-07 05:37:00.005392] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:26.869 [2024-12-07 05:37:00.005445] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:26.869 BaseBdev4 00:09:26.869 [2024-12-07 05:37:00.005767] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:26.869 [2024-12-07 05:37:00.005900] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:26.869 [2024-12-07 05:37:00.005956] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:26.869 [2024-12-07 05:37:00.006123] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.869 [ 00:09:26.869 { 00:09:26.869 "name": "BaseBdev4", 00:09:26.869 "aliases": [ 00:09:26.869 "2df4e2dc-892d-4c16-8863-184f64bf20a2" 00:09:26.869 ], 00:09:26.869 "product_name": "Malloc disk", 00:09:26.869 "block_size": 512, 00:09:26.869 "num_blocks": 65536, 00:09:26.869 "uuid": "2df4e2dc-892d-4c16-8863-184f64bf20a2", 00:09:26.869 "assigned_rate_limits": { 00:09:26.869 "rw_ios_per_sec": 0, 00:09:26.869 "rw_mbytes_per_sec": 0, 00:09:26.869 "r_mbytes_per_sec": 0, 00:09:26.869 "w_mbytes_per_sec": 0 00:09:26.869 }, 00:09:26.869 "claimed": true, 00:09:26.869 "claim_type": "exclusive_write", 00:09:26.869 "zoned": false, 00:09:26.869 "supported_io_types": { 00:09:26.869 "read": true, 00:09:26.869 "write": true, 00:09:26.869 "unmap": true, 00:09:26.869 "flush": true, 00:09:26.869 "reset": true, 00:09:26.869 "nvme_admin": false, 00:09:26.869 "nvme_io": false, 00:09:26.869 "nvme_io_md": false, 00:09:26.869 "write_zeroes": true, 00:09:26.869 "zcopy": true, 00:09:26.869 "get_zone_info": false, 00:09:26.869 "zone_management": false, 00:09:26.869 "zone_append": false, 00:09:26.869 "compare": false, 00:09:26.869 "compare_and_write": false, 00:09:26.869 "abort": true, 00:09:26.869 "seek_hole": false, 00:09:26.869 "seek_data": false, 00:09:26.869 "copy": true, 00:09:26.869 "nvme_iov_md": false 00:09:26.869 }, 00:09:26.869 "memory_domains": [ 00:09:26.869 { 00:09:26.869 "dma_device_id": "system", 00:09:26.869 "dma_device_type": 1 00:09:26.869 }, 00:09:26.869 { 00:09:26.869 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:26.869 "dma_device_type": 2 00:09:26.869 } 00:09:26.869 ], 00:09:26.869 "driver_specific": {} 00:09:26.869 } 00:09:26.869 ] 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:26.869 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.870 "name": "Existed_Raid", 00:09:26.870 "uuid": "e32e24c6-3a2e-47e4-b02c-be3732706575", 00:09:26.870 "strip_size_kb": 64, 00:09:26.870 "state": "online", 00:09:26.870 "raid_level": "raid0", 00:09:26.870 "superblock": true, 00:09:26.870 "num_base_bdevs": 4, 00:09:26.870 "num_base_bdevs_discovered": 4, 00:09:26.870 "num_base_bdevs_operational": 4, 00:09:26.870 "base_bdevs_list": [ 00:09:26.870 { 00:09:26.870 "name": "BaseBdev1", 00:09:26.870 "uuid": "82564374-df7b-477f-9da3-4176f8d34a9c", 00:09:26.870 "is_configured": true, 00:09:26.870 "data_offset": 2048, 00:09:26.870 "data_size": 63488 00:09:26.870 }, 00:09:26.870 { 00:09:26.870 "name": "BaseBdev2", 00:09:26.870 "uuid": "e26fd56a-f928-42bd-b17d-905f5ca082d2", 00:09:26.870 "is_configured": true, 00:09:26.870 "data_offset": 2048, 00:09:26.870 "data_size": 63488 00:09:26.870 }, 00:09:26.870 { 00:09:26.870 "name": "BaseBdev3", 00:09:26.870 "uuid": "ed92f78c-6775-4eeb-b44f-0a2be315aa16", 00:09:26.870 "is_configured": true, 00:09:26.870 "data_offset": 2048, 00:09:26.870 "data_size": 63488 00:09:26.870 }, 00:09:26.870 { 00:09:26.870 "name": "BaseBdev4", 00:09:26.870 "uuid": "2df4e2dc-892d-4c16-8863-184f64bf20a2", 00:09:26.870 "is_configured": true, 00:09:26.870 "data_offset": 2048, 00:09:26.870 "data_size": 63488 00:09:26.870 } 00:09:26.870 ] 00:09:26.870 }' 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.870 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.438 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:27.438 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:27.438 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:27.438 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:27.438 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.439 [2024-12-07 05:37:00.508656] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:27.439 "name": "Existed_Raid", 00:09:27.439 "aliases": [ 00:09:27.439 "e32e24c6-3a2e-47e4-b02c-be3732706575" 00:09:27.439 ], 00:09:27.439 "product_name": "Raid Volume", 00:09:27.439 "block_size": 512, 00:09:27.439 "num_blocks": 253952, 00:09:27.439 "uuid": "e32e24c6-3a2e-47e4-b02c-be3732706575", 00:09:27.439 "assigned_rate_limits": { 00:09:27.439 "rw_ios_per_sec": 0, 00:09:27.439 "rw_mbytes_per_sec": 0, 00:09:27.439 "r_mbytes_per_sec": 0, 00:09:27.439 "w_mbytes_per_sec": 0 00:09:27.439 }, 00:09:27.439 "claimed": false, 00:09:27.439 "zoned": false, 00:09:27.439 "supported_io_types": { 00:09:27.439 "read": true, 00:09:27.439 "write": true, 00:09:27.439 "unmap": true, 00:09:27.439 "flush": true, 00:09:27.439 "reset": true, 00:09:27.439 "nvme_admin": false, 00:09:27.439 "nvme_io": false, 00:09:27.439 "nvme_io_md": false, 00:09:27.439 "write_zeroes": true, 00:09:27.439 "zcopy": false, 00:09:27.439 "get_zone_info": false, 00:09:27.439 "zone_management": false, 00:09:27.439 "zone_append": false, 00:09:27.439 "compare": false, 00:09:27.439 "compare_and_write": false, 00:09:27.439 "abort": false, 00:09:27.439 "seek_hole": false, 00:09:27.439 "seek_data": false, 00:09:27.439 "copy": false, 00:09:27.439 "nvme_iov_md": false 00:09:27.439 }, 00:09:27.439 "memory_domains": [ 00:09:27.439 { 00:09:27.439 "dma_device_id": "system", 00:09:27.439 "dma_device_type": 1 00:09:27.439 }, 00:09:27.439 { 00:09:27.439 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.439 "dma_device_type": 2 00:09:27.439 }, 00:09:27.439 { 00:09:27.439 "dma_device_id": "system", 00:09:27.439 "dma_device_type": 1 00:09:27.439 }, 00:09:27.439 { 00:09:27.439 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.439 "dma_device_type": 2 00:09:27.439 }, 00:09:27.439 { 00:09:27.439 "dma_device_id": "system", 00:09:27.439 "dma_device_type": 1 00:09:27.439 }, 00:09:27.439 { 00:09:27.439 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.439 "dma_device_type": 2 00:09:27.439 }, 00:09:27.439 { 00:09:27.439 "dma_device_id": "system", 00:09:27.439 "dma_device_type": 1 00:09:27.439 }, 00:09:27.439 { 00:09:27.439 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.439 "dma_device_type": 2 00:09:27.439 } 00:09:27.439 ], 00:09:27.439 "driver_specific": { 00:09:27.439 "raid": { 00:09:27.439 "uuid": "e32e24c6-3a2e-47e4-b02c-be3732706575", 00:09:27.439 "strip_size_kb": 64, 00:09:27.439 "state": "online", 00:09:27.439 "raid_level": "raid0", 00:09:27.439 "superblock": true, 00:09:27.439 "num_base_bdevs": 4, 00:09:27.439 "num_base_bdevs_discovered": 4, 00:09:27.439 "num_base_bdevs_operational": 4, 00:09:27.439 "base_bdevs_list": [ 00:09:27.439 { 00:09:27.439 "name": "BaseBdev1", 00:09:27.439 "uuid": "82564374-df7b-477f-9da3-4176f8d34a9c", 00:09:27.439 "is_configured": true, 00:09:27.439 "data_offset": 2048, 00:09:27.439 "data_size": 63488 00:09:27.439 }, 00:09:27.439 { 00:09:27.439 "name": "BaseBdev2", 00:09:27.439 "uuid": "e26fd56a-f928-42bd-b17d-905f5ca082d2", 00:09:27.439 "is_configured": true, 00:09:27.439 "data_offset": 2048, 00:09:27.439 "data_size": 63488 00:09:27.439 }, 00:09:27.439 { 00:09:27.439 "name": "BaseBdev3", 00:09:27.439 "uuid": "ed92f78c-6775-4eeb-b44f-0a2be315aa16", 00:09:27.439 "is_configured": true, 00:09:27.439 "data_offset": 2048, 00:09:27.439 "data_size": 63488 00:09:27.439 }, 00:09:27.439 { 00:09:27.439 "name": "BaseBdev4", 00:09:27.439 "uuid": "2df4e2dc-892d-4c16-8863-184f64bf20a2", 00:09:27.439 "is_configured": true, 00:09:27.439 "data_offset": 2048, 00:09:27.439 "data_size": 63488 00:09:27.439 } 00:09:27.439 ] 00:09:27.439 } 00:09:27.439 } 00:09:27.439 }' 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:27.439 BaseBdev2 00:09:27.439 BaseBdev3 00:09:27.439 BaseBdev4' 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.439 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.699 [2024-12-07 05:37:00.839787] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:27.699 [2024-12-07 05:37:00.839861] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:27.699 [2024-12-07 05:37:00.839923] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.699 "name": "Existed_Raid", 00:09:27.699 "uuid": "e32e24c6-3a2e-47e4-b02c-be3732706575", 00:09:27.699 "strip_size_kb": 64, 00:09:27.699 "state": "offline", 00:09:27.699 "raid_level": "raid0", 00:09:27.699 "superblock": true, 00:09:27.699 "num_base_bdevs": 4, 00:09:27.699 "num_base_bdevs_discovered": 3, 00:09:27.699 "num_base_bdevs_operational": 3, 00:09:27.699 "base_bdevs_list": [ 00:09:27.699 { 00:09:27.699 "name": null, 00:09:27.699 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.699 "is_configured": false, 00:09:27.699 "data_offset": 0, 00:09:27.699 "data_size": 63488 00:09:27.699 }, 00:09:27.699 { 00:09:27.699 "name": "BaseBdev2", 00:09:27.699 "uuid": "e26fd56a-f928-42bd-b17d-905f5ca082d2", 00:09:27.699 "is_configured": true, 00:09:27.699 "data_offset": 2048, 00:09:27.699 "data_size": 63488 00:09:27.699 }, 00:09:27.699 { 00:09:27.699 "name": "BaseBdev3", 00:09:27.699 "uuid": "ed92f78c-6775-4eeb-b44f-0a2be315aa16", 00:09:27.699 "is_configured": true, 00:09:27.699 "data_offset": 2048, 00:09:27.699 "data_size": 63488 00:09:27.699 }, 00:09:27.699 { 00:09:27.699 "name": "BaseBdev4", 00:09:27.699 "uuid": "2df4e2dc-892d-4c16-8863-184f64bf20a2", 00:09:27.699 "is_configured": true, 00:09:27.699 "data_offset": 2048, 00:09:27.699 "data_size": 63488 00:09:27.699 } 00:09:27.699 ] 00:09:27.699 }' 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.699 05:37:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.959 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:27.959 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:27.959 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:27.959 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.959 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.959 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.959 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.219 [2024-12-07 05:37:01.338297] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.219 [2024-12-07 05:37:01.409559] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.219 [2024-12-07 05:37:01.480791] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:28.219 [2024-12-07 05:37:01.480835] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.219 BaseBdev2 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.219 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.479 [ 00:09:28.479 { 00:09:28.479 "name": "BaseBdev2", 00:09:28.479 "aliases": [ 00:09:28.479 "50b7bc51-504f-4a0c-befa-fac0a8f5375f" 00:09:28.479 ], 00:09:28.479 "product_name": "Malloc disk", 00:09:28.479 "block_size": 512, 00:09:28.479 "num_blocks": 65536, 00:09:28.479 "uuid": "50b7bc51-504f-4a0c-befa-fac0a8f5375f", 00:09:28.479 "assigned_rate_limits": { 00:09:28.479 "rw_ios_per_sec": 0, 00:09:28.479 "rw_mbytes_per_sec": 0, 00:09:28.479 "r_mbytes_per_sec": 0, 00:09:28.479 "w_mbytes_per_sec": 0 00:09:28.479 }, 00:09:28.479 "claimed": false, 00:09:28.479 "zoned": false, 00:09:28.479 "supported_io_types": { 00:09:28.479 "read": true, 00:09:28.479 "write": true, 00:09:28.479 "unmap": true, 00:09:28.479 "flush": true, 00:09:28.479 "reset": true, 00:09:28.479 "nvme_admin": false, 00:09:28.479 "nvme_io": false, 00:09:28.479 "nvme_io_md": false, 00:09:28.479 "write_zeroes": true, 00:09:28.479 "zcopy": true, 00:09:28.479 "get_zone_info": false, 00:09:28.479 "zone_management": false, 00:09:28.479 "zone_append": false, 00:09:28.479 "compare": false, 00:09:28.479 "compare_and_write": false, 00:09:28.479 "abort": true, 00:09:28.479 "seek_hole": false, 00:09:28.479 "seek_data": false, 00:09:28.479 "copy": true, 00:09:28.479 "nvme_iov_md": false 00:09:28.479 }, 00:09:28.479 "memory_domains": [ 00:09:28.479 { 00:09:28.479 "dma_device_id": "system", 00:09:28.479 "dma_device_type": 1 00:09:28.479 }, 00:09:28.479 { 00:09:28.479 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.479 "dma_device_type": 2 00:09:28.479 } 00:09:28.479 ], 00:09:28.479 "driver_specific": {} 00:09:28.479 } 00:09:28.479 ] 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.479 BaseBdev3 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.479 [ 00:09:28.479 { 00:09:28.479 "name": "BaseBdev3", 00:09:28.479 "aliases": [ 00:09:28.479 "0c219fd9-b0e6-4685-ad36-bda6c43c350e" 00:09:28.479 ], 00:09:28.479 "product_name": "Malloc disk", 00:09:28.479 "block_size": 512, 00:09:28.479 "num_blocks": 65536, 00:09:28.479 "uuid": "0c219fd9-b0e6-4685-ad36-bda6c43c350e", 00:09:28.479 "assigned_rate_limits": { 00:09:28.479 "rw_ios_per_sec": 0, 00:09:28.479 "rw_mbytes_per_sec": 0, 00:09:28.479 "r_mbytes_per_sec": 0, 00:09:28.479 "w_mbytes_per_sec": 0 00:09:28.479 }, 00:09:28.479 "claimed": false, 00:09:28.479 "zoned": false, 00:09:28.479 "supported_io_types": { 00:09:28.479 "read": true, 00:09:28.479 "write": true, 00:09:28.479 "unmap": true, 00:09:28.479 "flush": true, 00:09:28.479 "reset": true, 00:09:28.479 "nvme_admin": false, 00:09:28.479 "nvme_io": false, 00:09:28.479 "nvme_io_md": false, 00:09:28.479 "write_zeroes": true, 00:09:28.479 "zcopy": true, 00:09:28.479 "get_zone_info": false, 00:09:28.479 "zone_management": false, 00:09:28.479 "zone_append": false, 00:09:28.479 "compare": false, 00:09:28.479 "compare_and_write": false, 00:09:28.479 "abort": true, 00:09:28.479 "seek_hole": false, 00:09:28.479 "seek_data": false, 00:09:28.479 "copy": true, 00:09:28.479 "nvme_iov_md": false 00:09:28.479 }, 00:09:28.479 "memory_domains": [ 00:09:28.479 { 00:09:28.479 "dma_device_id": "system", 00:09:28.479 "dma_device_type": 1 00:09:28.479 }, 00:09:28.479 { 00:09:28.479 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.479 "dma_device_type": 2 00:09:28.479 } 00:09:28.479 ], 00:09:28.479 "driver_specific": {} 00:09:28.479 } 00:09:28.479 ] 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.479 BaseBdev4 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.479 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.480 [ 00:09:28.480 { 00:09:28.480 "name": "BaseBdev4", 00:09:28.480 "aliases": [ 00:09:28.480 "f4e4faa8-022d-4cfe-b349-1bb109b03cc4" 00:09:28.480 ], 00:09:28.480 "product_name": "Malloc disk", 00:09:28.480 "block_size": 512, 00:09:28.480 "num_blocks": 65536, 00:09:28.480 "uuid": "f4e4faa8-022d-4cfe-b349-1bb109b03cc4", 00:09:28.480 "assigned_rate_limits": { 00:09:28.480 "rw_ios_per_sec": 0, 00:09:28.480 "rw_mbytes_per_sec": 0, 00:09:28.480 "r_mbytes_per_sec": 0, 00:09:28.480 "w_mbytes_per_sec": 0 00:09:28.480 }, 00:09:28.480 "claimed": false, 00:09:28.480 "zoned": false, 00:09:28.480 "supported_io_types": { 00:09:28.480 "read": true, 00:09:28.480 "write": true, 00:09:28.480 "unmap": true, 00:09:28.480 "flush": true, 00:09:28.480 "reset": true, 00:09:28.480 "nvme_admin": false, 00:09:28.480 "nvme_io": false, 00:09:28.480 "nvme_io_md": false, 00:09:28.480 "write_zeroes": true, 00:09:28.480 "zcopy": true, 00:09:28.480 "get_zone_info": false, 00:09:28.480 "zone_management": false, 00:09:28.480 "zone_append": false, 00:09:28.480 "compare": false, 00:09:28.480 "compare_and_write": false, 00:09:28.480 "abort": true, 00:09:28.480 "seek_hole": false, 00:09:28.480 "seek_data": false, 00:09:28.480 "copy": true, 00:09:28.480 "nvme_iov_md": false 00:09:28.480 }, 00:09:28.480 "memory_domains": [ 00:09:28.480 { 00:09:28.480 "dma_device_id": "system", 00:09:28.480 "dma_device_type": 1 00:09:28.480 }, 00:09:28.480 { 00:09:28.480 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.480 "dma_device_type": 2 00:09:28.480 } 00:09:28.480 ], 00:09:28.480 "driver_specific": {} 00:09:28.480 } 00:09:28.480 ] 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.480 [2024-12-07 05:37:01.713868] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:28.480 [2024-12-07 05:37:01.713955] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:28.480 [2024-12-07 05:37:01.714020] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:28.480 [2024-12-07 05:37:01.715966] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:28.480 [2024-12-07 05:37:01.716066] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.480 "name": "Existed_Raid", 00:09:28.480 "uuid": "213a8586-429e-4447-b4ac-ca4010b000a0", 00:09:28.480 "strip_size_kb": 64, 00:09:28.480 "state": "configuring", 00:09:28.480 "raid_level": "raid0", 00:09:28.480 "superblock": true, 00:09:28.480 "num_base_bdevs": 4, 00:09:28.480 "num_base_bdevs_discovered": 3, 00:09:28.480 "num_base_bdevs_operational": 4, 00:09:28.480 "base_bdevs_list": [ 00:09:28.480 { 00:09:28.480 "name": "BaseBdev1", 00:09:28.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.480 "is_configured": false, 00:09:28.480 "data_offset": 0, 00:09:28.480 "data_size": 0 00:09:28.480 }, 00:09:28.480 { 00:09:28.480 "name": "BaseBdev2", 00:09:28.480 "uuid": "50b7bc51-504f-4a0c-befa-fac0a8f5375f", 00:09:28.480 "is_configured": true, 00:09:28.480 "data_offset": 2048, 00:09:28.480 "data_size": 63488 00:09:28.480 }, 00:09:28.480 { 00:09:28.480 "name": "BaseBdev3", 00:09:28.480 "uuid": "0c219fd9-b0e6-4685-ad36-bda6c43c350e", 00:09:28.480 "is_configured": true, 00:09:28.480 "data_offset": 2048, 00:09:28.480 "data_size": 63488 00:09:28.480 }, 00:09:28.480 { 00:09:28.480 "name": "BaseBdev4", 00:09:28.480 "uuid": "f4e4faa8-022d-4cfe-b349-1bb109b03cc4", 00:09:28.480 "is_configured": true, 00:09:28.480 "data_offset": 2048, 00:09:28.480 "data_size": 63488 00:09:28.480 } 00:09:28.480 ] 00:09:28.480 }' 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.480 05:37:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.048 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:29.048 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.048 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.048 [2024-12-07 05:37:02.181068] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:29.048 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.048 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:29.048 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.049 "name": "Existed_Raid", 00:09:29.049 "uuid": "213a8586-429e-4447-b4ac-ca4010b000a0", 00:09:29.049 "strip_size_kb": 64, 00:09:29.049 "state": "configuring", 00:09:29.049 "raid_level": "raid0", 00:09:29.049 "superblock": true, 00:09:29.049 "num_base_bdevs": 4, 00:09:29.049 "num_base_bdevs_discovered": 2, 00:09:29.049 "num_base_bdevs_operational": 4, 00:09:29.049 "base_bdevs_list": [ 00:09:29.049 { 00:09:29.049 "name": "BaseBdev1", 00:09:29.049 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.049 "is_configured": false, 00:09:29.049 "data_offset": 0, 00:09:29.049 "data_size": 0 00:09:29.049 }, 00:09:29.049 { 00:09:29.049 "name": null, 00:09:29.049 "uuid": "50b7bc51-504f-4a0c-befa-fac0a8f5375f", 00:09:29.049 "is_configured": false, 00:09:29.049 "data_offset": 0, 00:09:29.049 "data_size": 63488 00:09:29.049 }, 00:09:29.049 { 00:09:29.049 "name": "BaseBdev3", 00:09:29.049 "uuid": "0c219fd9-b0e6-4685-ad36-bda6c43c350e", 00:09:29.049 "is_configured": true, 00:09:29.049 "data_offset": 2048, 00:09:29.049 "data_size": 63488 00:09:29.049 }, 00:09:29.049 { 00:09:29.049 "name": "BaseBdev4", 00:09:29.049 "uuid": "f4e4faa8-022d-4cfe-b349-1bb109b03cc4", 00:09:29.049 "is_configured": true, 00:09:29.049 "data_offset": 2048, 00:09:29.049 "data_size": 63488 00:09:29.049 } 00:09:29.049 ] 00:09:29.049 }' 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.049 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.308 [2024-12-07 05:37:02.647244] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:29.308 BaseBdev1 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.308 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.308 [ 00:09:29.568 { 00:09:29.568 "name": "BaseBdev1", 00:09:29.568 "aliases": [ 00:09:29.568 "44e3f09c-c7c5-4e8f-a0f1-383e062de747" 00:09:29.568 ], 00:09:29.568 "product_name": "Malloc disk", 00:09:29.568 "block_size": 512, 00:09:29.568 "num_blocks": 65536, 00:09:29.568 "uuid": "44e3f09c-c7c5-4e8f-a0f1-383e062de747", 00:09:29.568 "assigned_rate_limits": { 00:09:29.568 "rw_ios_per_sec": 0, 00:09:29.568 "rw_mbytes_per_sec": 0, 00:09:29.568 "r_mbytes_per_sec": 0, 00:09:29.568 "w_mbytes_per_sec": 0 00:09:29.568 }, 00:09:29.568 "claimed": true, 00:09:29.568 "claim_type": "exclusive_write", 00:09:29.568 "zoned": false, 00:09:29.568 "supported_io_types": { 00:09:29.568 "read": true, 00:09:29.568 "write": true, 00:09:29.568 "unmap": true, 00:09:29.568 "flush": true, 00:09:29.568 "reset": true, 00:09:29.568 "nvme_admin": false, 00:09:29.568 "nvme_io": false, 00:09:29.568 "nvme_io_md": false, 00:09:29.568 "write_zeroes": true, 00:09:29.568 "zcopy": true, 00:09:29.568 "get_zone_info": false, 00:09:29.568 "zone_management": false, 00:09:29.568 "zone_append": false, 00:09:29.568 "compare": false, 00:09:29.568 "compare_and_write": false, 00:09:29.568 "abort": true, 00:09:29.568 "seek_hole": false, 00:09:29.568 "seek_data": false, 00:09:29.568 "copy": true, 00:09:29.568 "nvme_iov_md": false 00:09:29.568 }, 00:09:29.568 "memory_domains": [ 00:09:29.568 { 00:09:29.568 "dma_device_id": "system", 00:09:29.568 "dma_device_type": 1 00:09:29.568 }, 00:09:29.568 { 00:09:29.568 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.568 "dma_device_type": 2 00:09:29.568 } 00:09:29.568 ], 00:09:29.568 "driver_specific": {} 00:09:29.568 } 00:09:29.568 ] 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.568 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.569 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.569 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.569 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.569 "name": "Existed_Raid", 00:09:29.569 "uuid": "213a8586-429e-4447-b4ac-ca4010b000a0", 00:09:29.569 "strip_size_kb": 64, 00:09:29.569 "state": "configuring", 00:09:29.569 "raid_level": "raid0", 00:09:29.569 "superblock": true, 00:09:29.569 "num_base_bdevs": 4, 00:09:29.569 "num_base_bdevs_discovered": 3, 00:09:29.569 "num_base_bdevs_operational": 4, 00:09:29.569 "base_bdevs_list": [ 00:09:29.569 { 00:09:29.569 "name": "BaseBdev1", 00:09:29.569 "uuid": "44e3f09c-c7c5-4e8f-a0f1-383e062de747", 00:09:29.569 "is_configured": true, 00:09:29.569 "data_offset": 2048, 00:09:29.569 "data_size": 63488 00:09:29.569 }, 00:09:29.569 { 00:09:29.569 "name": null, 00:09:29.569 "uuid": "50b7bc51-504f-4a0c-befa-fac0a8f5375f", 00:09:29.569 "is_configured": false, 00:09:29.569 "data_offset": 0, 00:09:29.569 "data_size": 63488 00:09:29.569 }, 00:09:29.569 { 00:09:29.569 "name": "BaseBdev3", 00:09:29.569 "uuid": "0c219fd9-b0e6-4685-ad36-bda6c43c350e", 00:09:29.569 "is_configured": true, 00:09:29.569 "data_offset": 2048, 00:09:29.569 "data_size": 63488 00:09:29.569 }, 00:09:29.569 { 00:09:29.569 "name": "BaseBdev4", 00:09:29.569 "uuid": "f4e4faa8-022d-4cfe-b349-1bb109b03cc4", 00:09:29.569 "is_configured": true, 00:09:29.569 "data_offset": 2048, 00:09:29.569 "data_size": 63488 00:09:29.569 } 00:09:29.569 ] 00:09:29.569 }' 00:09:29.569 05:37:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.569 05:37:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.829 [2024-12-07 05:37:03.138600] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.829 "name": "Existed_Raid", 00:09:29.829 "uuid": "213a8586-429e-4447-b4ac-ca4010b000a0", 00:09:29.829 "strip_size_kb": 64, 00:09:29.829 "state": "configuring", 00:09:29.829 "raid_level": "raid0", 00:09:29.829 "superblock": true, 00:09:29.829 "num_base_bdevs": 4, 00:09:29.829 "num_base_bdevs_discovered": 2, 00:09:29.829 "num_base_bdevs_operational": 4, 00:09:29.829 "base_bdevs_list": [ 00:09:29.829 { 00:09:29.829 "name": "BaseBdev1", 00:09:29.829 "uuid": "44e3f09c-c7c5-4e8f-a0f1-383e062de747", 00:09:29.829 "is_configured": true, 00:09:29.829 "data_offset": 2048, 00:09:29.829 "data_size": 63488 00:09:29.829 }, 00:09:29.829 { 00:09:29.829 "name": null, 00:09:29.829 "uuid": "50b7bc51-504f-4a0c-befa-fac0a8f5375f", 00:09:29.829 "is_configured": false, 00:09:29.829 "data_offset": 0, 00:09:29.829 "data_size": 63488 00:09:29.829 }, 00:09:29.829 { 00:09:29.829 "name": null, 00:09:29.829 "uuid": "0c219fd9-b0e6-4685-ad36-bda6c43c350e", 00:09:29.829 "is_configured": false, 00:09:29.829 "data_offset": 0, 00:09:29.829 "data_size": 63488 00:09:29.829 }, 00:09:29.829 { 00:09:29.829 "name": "BaseBdev4", 00:09:29.829 "uuid": "f4e4faa8-022d-4cfe-b349-1bb109b03cc4", 00:09:29.829 "is_configured": true, 00:09:29.829 "data_offset": 2048, 00:09:29.829 "data_size": 63488 00:09:29.829 } 00:09:29.829 ] 00:09:29.829 }' 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.829 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.399 [2024-12-07 05:37:03.573882] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:30.399 "name": "Existed_Raid", 00:09:30.399 "uuid": "213a8586-429e-4447-b4ac-ca4010b000a0", 00:09:30.399 "strip_size_kb": 64, 00:09:30.399 "state": "configuring", 00:09:30.399 "raid_level": "raid0", 00:09:30.399 "superblock": true, 00:09:30.399 "num_base_bdevs": 4, 00:09:30.399 "num_base_bdevs_discovered": 3, 00:09:30.399 "num_base_bdevs_operational": 4, 00:09:30.399 "base_bdevs_list": [ 00:09:30.399 { 00:09:30.399 "name": "BaseBdev1", 00:09:30.399 "uuid": "44e3f09c-c7c5-4e8f-a0f1-383e062de747", 00:09:30.399 "is_configured": true, 00:09:30.399 "data_offset": 2048, 00:09:30.399 "data_size": 63488 00:09:30.399 }, 00:09:30.399 { 00:09:30.399 "name": null, 00:09:30.399 "uuid": "50b7bc51-504f-4a0c-befa-fac0a8f5375f", 00:09:30.399 "is_configured": false, 00:09:30.399 "data_offset": 0, 00:09:30.399 "data_size": 63488 00:09:30.399 }, 00:09:30.399 { 00:09:30.399 "name": "BaseBdev3", 00:09:30.399 "uuid": "0c219fd9-b0e6-4685-ad36-bda6c43c350e", 00:09:30.399 "is_configured": true, 00:09:30.399 "data_offset": 2048, 00:09:30.399 "data_size": 63488 00:09:30.399 }, 00:09:30.399 { 00:09:30.399 "name": "BaseBdev4", 00:09:30.399 "uuid": "f4e4faa8-022d-4cfe-b349-1bb109b03cc4", 00:09:30.399 "is_configured": true, 00:09:30.399 "data_offset": 2048, 00:09:30.399 "data_size": 63488 00:09:30.399 } 00:09:30.399 ] 00:09:30.399 }' 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:30.399 05:37:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.659 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:30.659 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.659 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.659 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.659 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.920 [2024-12-07 05:37:04.045168] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:30.920 "name": "Existed_Raid", 00:09:30.920 "uuid": "213a8586-429e-4447-b4ac-ca4010b000a0", 00:09:30.920 "strip_size_kb": 64, 00:09:30.920 "state": "configuring", 00:09:30.920 "raid_level": "raid0", 00:09:30.920 "superblock": true, 00:09:30.920 "num_base_bdevs": 4, 00:09:30.920 "num_base_bdevs_discovered": 2, 00:09:30.920 "num_base_bdevs_operational": 4, 00:09:30.920 "base_bdevs_list": [ 00:09:30.920 { 00:09:30.920 "name": null, 00:09:30.920 "uuid": "44e3f09c-c7c5-4e8f-a0f1-383e062de747", 00:09:30.920 "is_configured": false, 00:09:30.920 "data_offset": 0, 00:09:30.920 "data_size": 63488 00:09:30.920 }, 00:09:30.920 { 00:09:30.920 "name": null, 00:09:30.920 "uuid": "50b7bc51-504f-4a0c-befa-fac0a8f5375f", 00:09:30.920 "is_configured": false, 00:09:30.920 "data_offset": 0, 00:09:30.920 "data_size": 63488 00:09:30.920 }, 00:09:30.920 { 00:09:30.920 "name": "BaseBdev3", 00:09:30.920 "uuid": "0c219fd9-b0e6-4685-ad36-bda6c43c350e", 00:09:30.920 "is_configured": true, 00:09:30.920 "data_offset": 2048, 00:09:30.920 "data_size": 63488 00:09:30.920 }, 00:09:30.920 { 00:09:30.920 "name": "BaseBdev4", 00:09:30.920 "uuid": "f4e4faa8-022d-4cfe-b349-1bb109b03cc4", 00:09:30.920 "is_configured": true, 00:09:30.920 "data_offset": 2048, 00:09:30.920 "data_size": 63488 00:09:30.920 } 00:09:30.920 ] 00:09:30.920 }' 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:30.920 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.181 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.181 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:31.181 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.181 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.181 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.440 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:31.440 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:31.440 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.440 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.441 [2024-12-07 05:37:04.554803] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.441 "name": "Existed_Raid", 00:09:31.441 "uuid": "213a8586-429e-4447-b4ac-ca4010b000a0", 00:09:31.441 "strip_size_kb": 64, 00:09:31.441 "state": "configuring", 00:09:31.441 "raid_level": "raid0", 00:09:31.441 "superblock": true, 00:09:31.441 "num_base_bdevs": 4, 00:09:31.441 "num_base_bdevs_discovered": 3, 00:09:31.441 "num_base_bdevs_operational": 4, 00:09:31.441 "base_bdevs_list": [ 00:09:31.441 { 00:09:31.441 "name": null, 00:09:31.441 "uuid": "44e3f09c-c7c5-4e8f-a0f1-383e062de747", 00:09:31.441 "is_configured": false, 00:09:31.441 "data_offset": 0, 00:09:31.441 "data_size": 63488 00:09:31.441 }, 00:09:31.441 { 00:09:31.441 "name": "BaseBdev2", 00:09:31.441 "uuid": "50b7bc51-504f-4a0c-befa-fac0a8f5375f", 00:09:31.441 "is_configured": true, 00:09:31.441 "data_offset": 2048, 00:09:31.441 "data_size": 63488 00:09:31.441 }, 00:09:31.441 { 00:09:31.441 "name": "BaseBdev3", 00:09:31.441 "uuid": "0c219fd9-b0e6-4685-ad36-bda6c43c350e", 00:09:31.441 "is_configured": true, 00:09:31.441 "data_offset": 2048, 00:09:31.441 "data_size": 63488 00:09:31.441 }, 00:09:31.441 { 00:09:31.441 "name": "BaseBdev4", 00:09:31.441 "uuid": "f4e4faa8-022d-4cfe-b349-1bb109b03cc4", 00:09:31.441 "is_configured": true, 00:09:31.441 "data_offset": 2048, 00:09:31.441 "data_size": 63488 00:09:31.441 } 00:09:31.441 ] 00:09:31.441 }' 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.441 05:37:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.700 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.700 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.700 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.700 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:31.700 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 44e3f09c-c7c5-4e8f-a0f1-383e062de747 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.958 [2024-12-07 05:37:05.140788] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:31.958 [2024-12-07 05:37:05.141042] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:31.958 [2024-12-07 05:37:05.141090] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:31.958 [2024-12-07 05:37:05.141363] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:31.958 NewBaseBdev 00:09:31.958 [2024-12-07 05:37:05.141509] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:31.958 [2024-12-07 05:37:05.141527] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:31.958 [2024-12-07 05:37:05.141646] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.958 [ 00:09:31.958 { 00:09:31.958 "name": "NewBaseBdev", 00:09:31.958 "aliases": [ 00:09:31.958 "44e3f09c-c7c5-4e8f-a0f1-383e062de747" 00:09:31.958 ], 00:09:31.958 "product_name": "Malloc disk", 00:09:31.958 "block_size": 512, 00:09:31.958 "num_blocks": 65536, 00:09:31.958 "uuid": "44e3f09c-c7c5-4e8f-a0f1-383e062de747", 00:09:31.958 "assigned_rate_limits": { 00:09:31.958 "rw_ios_per_sec": 0, 00:09:31.958 "rw_mbytes_per_sec": 0, 00:09:31.958 "r_mbytes_per_sec": 0, 00:09:31.958 "w_mbytes_per_sec": 0 00:09:31.958 }, 00:09:31.958 "claimed": true, 00:09:31.958 "claim_type": "exclusive_write", 00:09:31.958 "zoned": false, 00:09:31.958 "supported_io_types": { 00:09:31.958 "read": true, 00:09:31.958 "write": true, 00:09:31.958 "unmap": true, 00:09:31.958 "flush": true, 00:09:31.958 "reset": true, 00:09:31.958 "nvme_admin": false, 00:09:31.958 "nvme_io": false, 00:09:31.958 "nvme_io_md": false, 00:09:31.958 "write_zeroes": true, 00:09:31.958 "zcopy": true, 00:09:31.958 "get_zone_info": false, 00:09:31.958 "zone_management": false, 00:09:31.958 "zone_append": false, 00:09:31.958 "compare": false, 00:09:31.958 "compare_and_write": false, 00:09:31.958 "abort": true, 00:09:31.958 "seek_hole": false, 00:09:31.958 "seek_data": false, 00:09:31.958 "copy": true, 00:09:31.958 "nvme_iov_md": false 00:09:31.958 }, 00:09:31.958 "memory_domains": [ 00:09:31.958 { 00:09:31.958 "dma_device_id": "system", 00:09:31.958 "dma_device_type": 1 00:09:31.958 }, 00:09:31.958 { 00:09:31.958 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:31.958 "dma_device_type": 2 00:09:31.958 } 00:09:31.958 ], 00:09:31.958 "driver_specific": {} 00:09:31.958 } 00:09:31.958 ] 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.958 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.959 "name": "Existed_Raid", 00:09:31.959 "uuid": "213a8586-429e-4447-b4ac-ca4010b000a0", 00:09:31.959 "strip_size_kb": 64, 00:09:31.959 "state": "online", 00:09:31.959 "raid_level": "raid0", 00:09:31.959 "superblock": true, 00:09:31.959 "num_base_bdevs": 4, 00:09:31.959 "num_base_bdevs_discovered": 4, 00:09:31.959 "num_base_bdevs_operational": 4, 00:09:31.959 "base_bdevs_list": [ 00:09:31.959 { 00:09:31.959 "name": "NewBaseBdev", 00:09:31.959 "uuid": "44e3f09c-c7c5-4e8f-a0f1-383e062de747", 00:09:31.959 "is_configured": true, 00:09:31.959 "data_offset": 2048, 00:09:31.959 "data_size": 63488 00:09:31.959 }, 00:09:31.959 { 00:09:31.959 "name": "BaseBdev2", 00:09:31.959 "uuid": "50b7bc51-504f-4a0c-befa-fac0a8f5375f", 00:09:31.959 "is_configured": true, 00:09:31.959 "data_offset": 2048, 00:09:31.959 "data_size": 63488 00:09:31.959 }, 00:09:31.959 { 00:09:31.959 "name": "BaseBdev3", 00:09:31.959 "uuid": "0c219fd9-b0e6-4685-ad36-bda6c43c350e", 00:09:31.959 "is_configured": true, 00:09:31.959 "data_offset": 2048, 00:09:31.959 "data_size": 63488 00:09:31.959 }, 00:09:31.959 { 00:09:31.959 "name": "BaseBdev4", 00:09:31.959 "uuid": "f4e4faa8-022d-4cfe-b349-1bb109b03cc4", 00:09:31.959 "is_configured": true, 00:09:31.959 "data_offset": 2048, 00:09:31.959 "data_size": 63488 00:09:31.959 } 00:09:31.959 ] 00:09:31.959 }' 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.959 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.526 [2024-12-07 05:37:05.612402] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.526 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:32.526 "name": "Existed_Raid", 00:09:32.526 "aliases": [ 00:09:32.526 "213a8586-429e-4447-b4ac-ca4010b000a0" 00:09:32.526 ], 00:09:32.526 "product_name": "Raid Volume", 00:09:32.526 "block_size": 512, 00:09:32.526 "num_blocks": 253952, 00:09:32.526 "uuid": "213a8586-429e-4447-b4ac-ca4010b000a0", 00:09:32.526 "assigned_rate_limits": { 00:09:32.526 "rw_ios_per_sec": 0, 00:09:32.526 "rw_mbytes_per_sec": 0, 00:09:32.526 "r_mbytes_per_sec": 0, 00:09:32.526 "w_mbytes_per_sec": 0 00:09:32.526 }, 00:09:32.526 "claimed": false, 00:09:32.526 "zoned": false, 00:09:32.526 "supported_io_types": { 00:09:32.526 "read": true, 00:09:32.526 "write": true, 00:09:32.526 "unmap": true, 00:09:32.526 "flush": true, 00:09:32.526 "reset": true, 00:09:32.526 "nvme_admin": false, 00:09:32.526 "nvme_io": false, 00:09:32.526 "nvme_io_md": false, 00:09:32.526 "write_zeroes": true, 00:09:32.526 "zcopy": false, 00:09:32.526 "get_zone_info": false, 00:09:32.526 "zone_management": false, 00:09:32.526 "zone_append": false, 00:09:32.526 "compare": false, 00:09:32.526 "compare_and_write": false, 00:09:32.526 "abort": false, 00:09:32.526 "seek_hole": false, 00:09:32.526 "seek_data": false, 00:09:32.526 "copy": false, 00:09:32.526 "nvme_iov_md": false 00:09:32.526 }, 00:09:32.526 "memory_domains": [ 00:09:32.526 { 00:09:32.526 "dma_device_id": "system", 00:09:32.526 "dma_device_type": 1 00:09:32.526 }, 00:09:32.526 { 00:09:32.526 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.526 "dma_device_type": 2 00:09:32.526 }, 00:09:32.526 { 00:09:32.526 "dma_device_id": "system", 00:09:32.526 "dma_device_type": 1 00:09:32.526 }, 00:09:32.526 { 00:09:32.526 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.526 "dma_device_type": 2 00:09:32.526 }, 00:09:32.526 { 00:09:32.526 "dma_device_id": "system", 00:09:32.526 "dma_device_type": 1 00:09:32.527 }, 00:09:32.527 { 00:09:32.527 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.527 "dma_device_type": 2 00:09:32.527 }, 00:09:32.527 { 00:09:32.527 "dma_device_id": "system", 00:09:32.527 "dma_device_type": 1 00:09:32.527 }, 00:09:32.527 { 00:09:32.527 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.527 "dma_device_type": 2 00:09:32.527 } 00:09:32.527 ], 00:09:32.527 "driver_specific": { 00:09:32.527 "raid": { 00:09:32.527 "uuid": "213a8586-429e-4447-b4ac-ca4010b000a0", 00:09:32.527 "strip_size_kb": 64, 00:09:32.527 "state": "online", 00:09:32.527 "raid_level": "raid0", 00:09:32.527 "superblock": true, 00:09:32.527 "num_base_bdevs": 4, 00:09:32.527 "num_base_bdevs_discovered": 4, 00:09:32.527 "num_base_bdevs_operational": 4, 00:09:32.527 "base_bdevs_list": [ 00:09:32.527 { 00:09:32.527 "name": "NewBaseBdev", 00:09:32.527 "uuid": "44e3f09c-c7c5-4e8f-a0f1-383e062de747", 00:09:32.527 "is_configured": true, 00:09:32.527 "data_offset": 2048, 00:09:32.527 "data_size": 63488 00:09:32.527 }, 00:09:32.527 { 00:09:32.527 "name": "BaseBdev2", 00:09:32.527 "uuid": "50b7bc51-504f-4a0c-befa-fac0a8f5375f", 00:09:32.527 "is_configured": true, 00:09:32.527 "data_offset": 2048, 00:09:32.527 "data_size": 63488 00:09:32.527 }, 00:09:32.527 { 00:09:32.527 "name": "BaseBdev3", 00:09:32.527 "uuid": "0c219fd9-b0e6-4685-ad36-bda6c43c350e", 00:09:32.527 "is_configured": true, 00:09:32.527 "data_offset": 2048, 00:09:32.527 "data_size": 63488 00:09:32.527 }, 00:09:32.527 { 00:09:32.527 "name": "BaseBdev4", 00:09:32.527 "uuid": "f4e4faa8-022d-4cfe-b349-1bb109b03cc4", 00:09:32.527 "is_configured": true, 00:09:32.527 "data_offset": 2048, 00:09:32.527 "data_size": 63488 00:09:32.527 } 00:09:32.527 ] 00:09:32.527 } 00:09:32.527 } 00:09:32.527 }' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:32.527 BaseBdev2 00:09:32.527 BaseBdev3 00:09:32.527 BaseBdev4' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:32.527 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.786 [2024-12-07 05:37:05.935541] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:32.786 [2024-12-07 05:37:05.935615] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:32.786 [2024-12-07 05:37:05.935719] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:32.786 [2024-12-07 05:37:05.935812] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:32.786 [2024-12-07 05:37:05.935826] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 80663 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 80663 ']' 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 80663 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80663 00:09:32.786 killing process with pid 80663 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80663' 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 80663 00:09:32.786 [2024-12-07 05:37:05.981275] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:32.786 05:37:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 80663 00:09:32.786 [2024-12-07 05:37:06.021280] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:33.045 05:37:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:33.046 00:09:33.046 real 0m9.567s 00:09:33.046 user 0m16.451s 00:09:33.046 sys 0m1.933s 00:09:33.046 05:37:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:33.046 ************************************ 00:09:33.046 END TEST raid_state_function_test_sb 00:09:33.046 ************************************ 00:09:33.046 05:37:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.046 05:37:06 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:09:33.046 05:37:06 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:33.046 05:37:06 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:33.046 05:37:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:33.046 ************************************ 00:09:33.046 START TEST raid_superblock_test 00:09:33.046 ************************************ 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 4 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:33.046 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=81311 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 81311 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 81311 ']' 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:33.046 05:37:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.046 [2024-12-07 05:37:06.392704] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:09:33.046 [2024-12-07 05:37:06.392921] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81311 ] 00:09:33.304 [2024-12-07 05:37:06.527357] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:33.304 [2024-12-07 05:37:06.552931] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:33.304 [2024-12-07 05:37:06.595379] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:33.304 [2024-12-07 05:37:06.595502] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.243 malloc1 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.243 [2024-12-07 05:37:07.274728] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:34.243 [2024-12-07 05:37:07.274858] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:34.243 [2024-12-07 05:37:07.274905] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:34.243 [2024-12-07 05:37:07.274944] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:34.243 [2024-12-07 05:37:07.277183] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:34.243 [2024-12-07 05:37:07.277261] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:34.243 pt1 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.243 malloc2 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.243 [2024-12-07 05:37:07.303212] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:34.243 [2024-12-07 05:37:07.303312] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:34.243 [2024-12-07 05:37:07.303358] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:34.243 [2024-12-07 05:37:07.303390] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:34.243 [2024-12-07 05:37:07.305499] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:34.243 [2024-12-07 05:37:07.305568] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:34.243 pt2 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.243 malloc3 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.243 [2024-12-07 05:37:07.335696] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:34.243 [2024-12-07 05:37:07.335788] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:34.243 [2024-12-07 05:37:07.335847] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:34.243 [2024-12-07 05:37:07.335891] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:34.243 [2024-12-07 05:37:07.338046] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:34.243 [2024-12-07 05:37:07.338117] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:34.243 pt3 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.243 malloc4 00:09:34.243 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.244 [2024-12-07 05:37:07.375771] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:34.244 [2024-12-07 05:37:07.375824] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:34.244 [2024-12-07 05:37:07.375839] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:34.244 [2024-12-07 05:37:07.375851] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:34.244 [2024-12-07 05:37:07.377947] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:34.244 [2024-12-07 05:37:07.378034] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:34.244 pt4 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.244 [2024-12-07 05:37:07.387784] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:34.244 [2024-12-07 05:37:07.389698] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:34.244 [2024-12-07 05:37:07.389765] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:34.244 [2024-12-07 05:37:07.389833] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:34.244 [2024-12-07 05:37:07.389986] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:34.244 [2024-12-07 05:37:07.389999] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:34.244 [2024-12-07 05:37:07.390253] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:34.244 [2024-12-07 05:37:07.390382] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:34.244 [2024-12-07 05:37:07.390391] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:34.244 [2024-12-07 05:37:07.390537] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.244 "name": "raid_bdev1", 00:09:34.244 "uuid": "0e672225-9f52-44e5-b3c1-8884fa5cb145", 00:09:34.244 "strip_size_kb": 64, 00:09:34.244 "state": "online", 00:09:34.244 "raid_level": "raid0", 00:09:34.244 "superblock": true, 00:09:34.244 "num_base_bdevs": 4, 00:09:34.244 "num_base_bdevs_discovered": 4, 00:09:34.244 "num_base_bdevs_operational": 4, 00:09:34.244 "base_bdevs_list": [ 00:09:34.244 { 00:09:34.244 "name": "pt1", 00:09:34.244 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:34.244 "is_configured": true, 00:09:34.244 "data_offset": 2048, 00:09:34.244 "data_size": 63488 00:09:34.244 }, 00:09:34.244 { 00:09:34.244 "name": "pt2", 00:09:34.244 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:34.244 "is_configured": true, 00:09:34.244 "data_offset": 2048, 00:09:34.244 "data_size": 63488 00:09:34.244 }, 00:09:34.244 { 00:09:34.244 "name": "pt3", 00:09:34.244 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:34.244 "is_configured": true, 00:09:34.244 "data_offset": 2048, 00:09:34.244 "data_size": 63488 00:09:34.244 }, 00:09:34.244 { 00:09:34.244 "name": "pt4", 00:09:34.244 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:34.244 "is_configured": true, 00:09:34.244 "data_offset": 2048, 00:09:34.244 "data_size": 63488 00:09:34.244 } 00:09:34.244 ] 00:09:34.244 }' 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.244 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.559 [2024-12-07 05:37:07.851319] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:34.559 "name": "raid_bdev1", 00:09:34.559 "aliases": [ 00:09:34.559 "0e672225-9f52-44e5-b3c1-8884fa5cb145" 00:09:34.559 ], 00:09:34.559 "product_name": "Raid Volume", 00:09:34.559 "block_size": 512, 00:09:34.559 "num_blocks": 253952, 00:09:34.559 "uuid": "0e672225-9f52-44e5-b3c1-8884fa5cb145", 00:09:34.559 "assigned_rate_limits": { 00:09:34.559 "rw_ios_per_sec": 0, 00:09:34.559 "rw_mbytes_per_sec": 0, 00:09:34.559 "r_mbytes_per_sec": 0, 00:09:34.559 "w_mbytes_per_sec": 0 00:09:34.559 }, 00:09:34.559 "claimed": false, 00:09:34.559 "zoned": false, 00:09:34.559 "supported_io_types": { 00:09:34.559 "read": true, 00:09:34.559 "write": true, 00:09:34.559 "unmap": true, 00:09:34.559 "flush": true, 00:09:34.559 "reset": true, 00:09:34.559 "nvme_admin": false, 00:09:34.559 "nvme_io": false, 00:09:34.559 "nvme_io_md": false, 00:09:34.559 "write_zeroes": true, 00:09:34.559 "zcopy": false, 00:09:34.559 "get_zone_info": false, 00:09:34.559 "zone_management": false, 00:09:34.559 "zone_append": false, 00:09:34.559 "compare": false, 00:09:34.559 "compare_and_write": false, 00:09:34.559 "abort": false, 00:09:34.559 "seek_hole": false, 00:09:34.559 "seek_data": false, 00:09:34.559 "copy": false, 00:09:34.559 "nvme_iov_md": false 00:09:34.559 }, 00:09:34.559 "memory_domains": [ 00:09:34.559 { 00:09:34.559 "dma_device_id": "system", 00:09:34.559 "dma_device_type": 1 00:09:34.559 }, 00:09:34.559 { 00:09:34.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.559 "dma_device_type": 2 00:09:34.559 }, 00:09:34.559 { 00:09:34.559 "dma_device_id": "system", 00:09:34.559 "dma_device_type": 1 00:09:34.559 }, 00:09:34.559 { 00:09:34.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.559 "dma_device_type": 2 00:09:34.559 }, 00:09:34.559 { 00:09:34.559 "dma_device_id": "system", 00:09:34.559 "dma_device_type": 1 00:09:34.559 }, 00:09:34.559 { 00:09:34.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.559 "dma_device_type": 2 00:09:34.559 }, 00:09:34.559 { 00:09:34.559 "dma_device_id": "system", 00:09:34.559 "dma_device_type": 1 00:09:34.559 }, 00:09:34.559 { 00:09:34.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.559 "dma_device_type": 2 00:09:34.559 } 00:09:34.559 ], 00:09:34.559 "driver_specific": { 00:09:34.559 "raid": { 00:09:34.559 "uuid": "0e672225-9f52-44e5-b3c1-8884fa5cb145", 00:09:34.559 "strip_size_kb": 64, 00:09:34.559 "state": "online", 00:09:34.559 "raid_level": "raid0", 00:09:34.559 "superblock": true, 00:09:34.559 "num_base_bdevs": 4, 00:09:34.559 "num_base_bdevs_discovered": 4, 00:09:34.559 "num_base_bdevs_operational": 4, 00:09:34.559 "base_bdevs_list": [ 00:09:34.559 { 00:09:34.559 "name": "pt1", 00:09:34.559 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:34.559 "is_configured": true, 00:09:34.559 "data_offset": 2048, 00:09:34.559 "data_size": 63488 00:09:34.559 }, 00:09:34.559 { 00:09:34.559 "name": "pt2", 00:09:34.559 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:34.559 "is_configured": true, 00:09:34.559 "data_offset": 2048, 00:09:34.559 "data_size": 63488 00:09:34.559 }, 00:09:34.559 { 00:09:34.559 "name": "pt3", 00:09:34.559 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:34.559 "is_configured": true, 00:09:34.559 "data_offset": 2048, 00:09:34.559 "data_size": 63488 00:09:34.559 }, 00:09:34.559 { 00:09:34.559 "name": "pt4", 00:09:34.559 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:34.559 "is_configured": true, 00:09:34.559 "data_offset": 2048, 00:09:34.559 "data_size": 63488 00:09:34.559 } 00:09:34.559 ] 00:09:34.559 } 00:09:34.559 } 00:09:34.559 }' 00:09:34.559 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:34.818 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:34.818 pt2 00:09:34.818 pt3 00:09:34.818 pt4' 00:09:34.818 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.818 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:34.818 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:34.818 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:34.818 05:37:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.818 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.818 05:37:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.818 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.819 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:34.819 [2024-12-07 05:37:08.174717] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=0e672225-9f52-44e5-b3c1-8884fa5cb145 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 0e672225-9f52-44e5-b3c1-8884fa5cb145 ']' 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.079 [2024-12-07 05:37:08.226314] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:35.079 [2024-12-07 05:37:08.226390] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:35.079 [2024-12-07 05:37:08.226515] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:35.079 [2024-12-07 05:37:08.226656] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:35.079 [2024-12-07 05:37:08.226716] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.079 [2024-12-07 05:37:08.362102] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:35.079 [2024-12-07 05:37:08.364099] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:35.079 [2024-12-07 05:37:08.364192] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:35.079 [2024-12-07 05:37:08.364251] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:09:35.079 [2024-12-07 05:37:08.364324] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:35.079 [2024-12-07 05:37:08.364407] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:35.079 [2024-12-07 05:37:08.364466] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:35.079 [2024-12-07 05:37:08.364521] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:09:35.079 [2024-12-07 05:37:08.364564] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:35.079 [2024-12-07 05:37:08.364576] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:35.079 request: 00:09:35.079 { 00:09:35.079 "name": "raid_bdev1", 00:09:35.079 "raid_level": "raid0", 00:09:35.079 "base_bdevs": [ 00:09:35.079 "malloc1", 00:09:35.079 "malloc2", 00:09:35.079 "malloc3", 00:09:35.079 "malloc4" 00:09:35.079 ], 00:09:35.079 "strip_size_kb": 64, 00:09:35.079 "superblock": false, 00:09:35.079 "method": "bdev_raid_create", 00:09:35.079 "req_id": 1 00:09:35.079 } 00:09:35.079 Got JSON-RPC error response 00:09:35.079 response: 00:09:35.079 { 00:09:35.079 "code": -17, 00:09:35.079 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:35.079 } 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.079 [2024-12-07 05:37:08.413985] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:35.079 [2024-12-07 05:37:08.414074] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:35.079 [2024-12-07 05:37:08.414122] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:35.079 [2024-12-07 05:37:08.414151] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:35.079 [2024-12-07 05:37:08.416482] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:35.079 [2024-12-07 05:37:08.416551] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:35.079 [2024-12-07 05:37:08.416677] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:35.079 [2024-12-07 05:37:08.416747] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:35.079 pt1 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:35.079 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.338 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:35.338 "name": "raid_bdev1", 00:09:35.338 "uuid": "0e672225-9f52-44e5-b3c1-8884fa5cb145", 00:09:35.338 "strip_size_kb": 64, 00:09:35.338 "state": "configuring", 00:09:35.338 "raid_level": "raid0", 00:09:35.338 "superblock": true, 00:09:35.338 "num_base_bdevs": 4, 00:09:35.338 "num_base_bdevs_discovered": 1, 00:09:35.338 "num_base_bdevs_operational": 4, 00:09:35.338 "base_bdevs_list": [ 00:09:35.338 { 00:09:35.338 "name": "pt1", 00:09:35.338 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:35.338 "is_configured": true, 00:09:35.338 "data_offset": 2048, 00:09:35.338 "data_size": 63488 00:09:35.338 }, 00:09:35.338 { 00:09:35.338 "name": null, 00:09:35.338 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:35.338 "is_configured": false, 00:09:35.338 "data_offset": 2048, 00:09:35.338 "data_size": 63488 00:09:35.338 }, 00:09:35.338 { 00:09:35.338 "name": null, 00:09:35.338 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:35.338 "is_configured": false, 00:09:35.338 "data_offset": 2048, 00:09:35.338 "data_size": 63488 00:09:35.338 }, 00:09:35.338 { 00:09:35.338 "name": null, 00:09:35.338 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:35.338 "is_configured": false, 00:09:35.338 "data_offset": 2048, 00:09:35.338 "data_size": 63488 00:09:35.338 } 00:09:35.338 ] 00:09:35.338 }' 00:09:35.338 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:35.338 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.598 [2024-12-07 05:37:08.845284] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:35.598 [2024-12-07 05:37:08.845364] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:35.598 [2024-12-07 05:37:08.845387] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:09:35.598 [2024-12-07 05:37:08.845398] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:35.598 [2024-12-07 05:37:08.845834] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:35.598 [2024-12-07 05:37:08.845854] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:35.598 [2024-12-07 05:37:08.845939] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:35.598 [2024-12-07 05:37:08.845961] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:35.598 pt2 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.598 [2024-12-07 05:37:08.857280] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.598 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:35.598 "name": "raid_bdev1", 00:09:35.598 "uuid": "0e672225-9f52-44e5-b3c1-8884fa5cb145", 00:09:35.598 "strip_size_kb": 64, 00:09:35.598 "state": "configuring", 00:09:35.598 "raid_level": "raid0", 00:09:35.598 "superblock": true, 00:09:35.598 "num_base_bdevs": 4, 00:09:35.598 "num_base_bdevs_discovered": 1, 00:09:35.598 "num_base_bdevs_operational": 4, 00:09:35.598 "base_bdevs_list": [ 00:09:35.598 { 00:09:35.598 "name": "pt1", 00:09:35.598 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:35.598 "is_configured": true, 00:09:35.598 "data_offset": 2048, 00:09:35.598 "data_size": 63488 00:09:35.598 }, 00:09:35.598 { 00:09:35.598 "name": null, 00:09:35.598 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:35.598 "is_configured": false, 00:09:35.598 "data_offset": 0, 00:09:35.598 "data_size": 63488 00:09:35.598 }, 00:09:35.599 { 00:09:35.599 "name": null, 00:09:35.599 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:35.599 "is_configured": false, 00:09:35.599 "data_offset": 2048, 00:09:35.599 "data_size": 63488 00:09:35.599 }, 00:09:35.599 { 00:09:35.599 "name": null, 00:09:35.599 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:35.599 "is_configured": false, 00:09:35.599 "data_offset": 2048, 00:09:35.599 "data_size": 63488 00:09:35.599 } 00:09:35.599 ] 00:09:35.599 }' 00:09:35.599 05:37:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:35.599 05:37:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.166 [2024-12-07 05:37:09.316504] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:36.166 [2024-12-07 05:37:09.316641] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:36.166 [2024-12-07 05:37:09.316665] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:36.166 [2024-12-07 05:37:09.316675] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:36.166 [2024-12-07 05:37:09.317075] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:36.166 [2024-12-07 05:37:09.317096] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:36.166 [2024-12-07 05:37:09.317187] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:36.166 [2024-12-07 05:37:09.317220] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:36.166 pt2 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.166 [2024-12-07 05:37:09.324449] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:36.166 [2024-12-07 05:37:09.324504] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:36.166 [2024-12-07 05:37:09.324520] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:36.166 [2024-12-07 05:37:09.324531] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:36.166 [2024-12-07 05:37:09.324903] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:36.166 [2024-12-07 05:37:09.324924] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:36.166 [2024-12-07 05:37:09.324978] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:36.166 [2024-12-07 05:37:09.324999] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:36.166 pt3 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:36.166 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.167 [2024-12-07 05:37:09.332434] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:36.167 [2024-12-07 05:37:09.332486] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:36.167 [2024-12-07 05:37:09.332499] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:09:36.167 [2024-12-07 05:37:09.332509] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:36.167 [2024-12-07 05:37:09.332817] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:36.167 [2024-12-07 05:37:09.332841] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:36.167 [2024-12-07 05:37:09.332893] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:09:36.167 [2024-12-07 05:37:09.332913] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:36.167 [2024-12-07 05:37:09.333007] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:36.167 [2024-12-07 05:37:09.333018] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:36.167 [2024-12-07 05:37:09.333258] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:36.167 [2024-12-07 05:37:09.333379] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:36.167 [2024-12-07 05:37:09.333387] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:36.167 [2024-12-07 05:37:09.333485] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:36.167 pt4 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.167 "name": "raid_bdev1", 00:09:36.167 "uuid": "0e672225-9f52-44e5-b3c1-8884fa5cb145", 00:09:36.167 "strip_size_kb": 64, 00:09:36.167 "state": "online", 00:09:36.167 "raid_level": "raid0", 00:09:36.167 "superblock": true, 00:09:36.167 "num_base_bdevs": 4, 00:09:36.167 "num_base_bdevs_discovered": 4, 00:09:36.167 "num_base_bdevs_operational": 4, 00:09:36.167 "base_bdevs_list": [ 00:09:36.167 { 00:09:36.167 "name": "pt1", 00:09:36.167 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:36.167 "is_configured": true, 00:09:36.167 "data_offset": 2048, 00:09:36.167 "data_size": 63488 00:09:36.167 }, 00:09:36.167 { 00:09:36.167 "name": "pt2", 00:09:36.167 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:36.167 "is_configured": true, 00:09:36.167 "data_offset": 2048, 00:09:36.167 "data_size": 63488 00:09:36.167 }, 00:09:36.167 { 00:09:36.167 "name": "pt3", 00:09:36.167 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:36.167 "is_configured": true, 00:09:36.167 "data_offset": 2048, 00:09:36.167 "data_size": 63488 00:09:36.167 }, 00:09:36.167 { 00:09:36.167 "name": "pt4", 00:09:36.167 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:36.167 "is_configured": true, 00:09:36.167 "data_offset": 2048, 00:09:36.167 "data_size": 63488 00:09:36.167 } 00:09:36.167 ] 00:09:36.167 }' 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.167 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.425 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:36.425 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:36.425 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:36.426 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:36.426 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:36.426 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:36.426 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:36.426 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:36.426 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.426 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.426 [2024-12-07 05:37:09.792047] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:36.685 "name": "raid_bdev1", 00:09:36.685 "aliases": [ 00:09:36.685 "0e672225-9f52-44e5-b3c1-8884fa5cb145" 00:09:36.685 ], 00:09:36.685 "product_name": "Raid Volume", 00:09:36.685 "block_size": 512, 00:09:36.685 "num_blocks": 253952, 00:09:36.685 "uuid": "0e672225-9f52-44e5-b3c1-8884fa5cb145", 00:09:36.685 "assigned_rate_limits": { 00:09:36.685 "rw_ios_per_sec": 0, 00:09:36.685 "rw_mbytes_per_sec": 0, 00:09:36.685 "r_mbytes_per_sec": 0, 00:09:36.685 "w_mbytes_per_sec": 0 00:09:36.685 }, 00:09:36.685 "claimed": false, 00:09:36.685 "zoned": false, 00:09:36.685 "supported_io_types": { 00:09:36.685 "read": true, 00:09:36.685 "write": true, 00:09:36.685 "unmap": true, 00:09:36.685 "flush": true, 00:09:36.685 "reset": true, 00:09:36.685 "nvme_admin": false, 00:09:36.685 "nvme_io": false, 00:09:36.685 "nvme_io_md": false, 00:09:36.685 "write_zeroes": true, 00:09:36.685 "zcopy": false, 00:09:36.685 "get_zone_info": false, 00:09:36.685 "zone_management": false, 00:09:36.685 "zone_append": false, 00:09:36.685 "compare": false, 00:09:36.685 "compare_and_write": false, 00:09:36.685 "abort": false, 00:09:36.685 "seek_hole": false, 00:09:36.685 "seek_data": false, 00:09:36.685 "copy": false, 00:09:36.685 "nvme_iov_md": false 00:09:36.685 }, 00:09:36.685 "memory_domains": [ 00:09:36.685 { 00:09:36.685 "dma_device_id": "system", 00:09:36.685 "dma_device_type": 1 00:09:36.685 }, 00:09:36.685 { 00:09:36.685 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.685 "dma_device_type": 2 00:09:36.685 }, 00:09:36.685 { 00:09:36.685 "dma_device_id": "system", 00:09:36.685 "dma_device_type": 1 00:09:36.685 }, 00:09:36.685 { 00:09:36.685 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.685 "dma_device_type": 2 00:09:36.685 }, 00:09:36.685 { 00:09:36.685 "dma_device_id": "system", 00:09:36.685 "dma_device_type": 1 00:09:36.685 }, 00:09:36.685 { 00:09:36.685 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.685 "dma_device_type": 2 00:09:36.685 }, 00:09:36.685 { 00:09:36.685 "dma_device_id": "system", 00:09:36.685 "dma_device_type": 1 00:09:36.685 }, 00:09:36.685 { 00:09:36.685 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.685 "dma_device_type": 2 00:09:36.685 } 00:09:36.685 ], 00:09:36.685 "driver_specific": { 00:09:36.685 "raid": { 00:09:36.685 "uuid": "0e672225-9f52-44e5-b3c1-8884fa5cb145", 00:09:36.685 "strip_size_kb": 64, 00:09:36.685 "state": "online", 00:09:36.685 "raid_level": "raid0", 00:09:36.685 "superblock": true, 00:09:36.685 "num_base_bdevs": 4, 00:09:36.685 "num_base_bdevs_discovered": 4, 00:09:36.685 "num_base_bdevs_operational": 4, 00:09:36.685 "base_bdevs_list": [ 00:09:36.685 { 00:09:36.685 "name": "pt1", 00:09:36.685 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:36.685 "is_configured": true, 00:09:36.685 "data_offset": 2048, 00:09:36.685 "data_size": 63488 00:09:36.685 }, 00:09:36.685 { 00:09:36.685 "name": "pt2", 00:09:36.685 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:36.685 "is_configured": true, 00:09:36.685 "data_offset": 2048, 00:09:36.685 "data_size": 63488 00:09:36.685 }, 00:09:36.685 { 00:09:36.685 "name": "pt3", 00:09:36.685 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:36.685 "is_configured": true, 00:09:36.685 "data_offset": 2048, 00:09:36.685 "data_size": 63488 00:09:36.685 }, 00:09:36.685 { 00:09:36.685 "name": "pt4", 00:09:36.685 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:36.685 "is_configured": true, 00:09:36.685 "data_offset": 2048, 00:09:36.685 "data_size": 63488 00:09:36.685 } 00:09:36.685 ] 00:09:36.685 } 00:09:36.685 } 00:09:36.685 }' 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:36.685 pt2 00:09:36.685 pt3 00:09:36.685 pt4' 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.685 05:37:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.685 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:36.685 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:36.685 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:36.685 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:36.685 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.686 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.686 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:36.686 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.945 [2024-12-07 05:37:10.115409] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 0e672225-9f52-44e5-b3c1-8884fa5cb145 '!=' 0e672225-9f52-44e5-b3c1-8884fa5cb145 ']' 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 81311 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 81311 ']' 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 81311 00:09:36.945 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:36.946 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:36.946 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81311 00:09:36.946 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:36.946 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:36.946 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81311' 00:09:36.946 killing process with pid 81311 00:09:36.946 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 81311 00:09:36.946 [2024-12-07 05:37:10.192067] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:36.946 [2024-12-07 05:37:10.192200] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:36.946 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 81311 00:09:36.946 [2024-12-07 05:37:10.192297] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:36.946 [2024-12-07 05:37:10.192311] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:36.946 [2024-12-07 05:37:10.235284] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:37.205 05:37:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:37.205 00:09:37.205 real 0m4.142s 00:09:37.205 user 0m6.586s 00:09:37.205 sys 0m0.871s 00:09:37.205 ************************************ 00:09:37.205 END TEST raid_superblock_test 00:09:37.205 ************************************ 00:09:37.205 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:37.205 05:37:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.205 05:37:10 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:09:37.205 05:37:10 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:37.205 05:37:10 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:37.205 05:37:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:37.205 ************************************ 00:09:37.205 START TEST raid_read_error_test 00:09:37.205 ************************************ 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 read 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.ltCeWtklGa 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81559 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81559 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 81559 ']' 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:37.205 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:37.205 05:37:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.464 [2024-12-07 05:37:10.614466] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:09:37.464 [2024-12-07 05:37:10.614615] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81559 ] 00:09:37.464 [2024-12-07 05:37:10.767985] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:37.464 [2024-12-07 05:37:10.793833] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:37.723 [2024-12-07 05:37:10.836993] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:37.723 [2024-12-07 05:37:10.837028] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.292 BaseBdev1_malloc 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.292 true 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.292 [2024-12-07 05:37:11.468598] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:38.292 [2024-12-07 05:37:11.468673] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:38.292 [2024-12-07 05:37:11.468696] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:38.292 [2024-12-07 05:37:11.468704] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:38.292 [2024-12-07 05:37:11.471078] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:38.292 [2024-12-07 05:37:11.471161] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:38.292 BaseBdev1 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.292 BaseBdev2_malloc 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.292 true 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:38.292 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.293 [2024-12-07 05:37:11.497532] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:38.293 [2024-12-07 05:37:11.497639] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:38.293 [2024-12-07 05:37:11.497677] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:38.293 [2024-12-07 05:37:11.497696] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:38.293 [2024-12-07 05:37:11.499844] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:38.293 [2024-12-07 05:37:11.499880] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:38.293 BaseBdev2 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.293 BaseBdev3_malloc 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.293 true 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.293 [2024-12-07 05:37:11.538135] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:38.293 [2024-12-07 05:37:11.538186] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:38.293 [2024-12-07 05:37:11.538205] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:38.293 [2024-12-07 05:37:11.538214] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:38.293 [2024-12-07 05:37:11.540340] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:38.293 [2024-12-07 05:37:11.540376] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:38.293 BaseBdev3 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.293 BaseBdev4_malloc 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.293 true 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.293 [2024-12-07 05:37:11.586729] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:09:38.293 [2024-12-07 05:37:11.586779] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:38.293 [2024-12-07 05:37:11.586819] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:38.293 [2024-12-07 05:37:11.586828] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:38.293 [2024-12-07 05:37:11.589121] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:38.293 [2024-12-07 05:37:11.589198] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:09:38.293 BaseBdev4 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.293 [2024-12-07 05:37:11.598757] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:38.293 [2024-12-07 05:37:11.600653] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:38.293 [2024-12-07 05:37:11.600734] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:38.293 [2024-12-07 05:37:11.600787] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:38.293 [2024-12-07 05:37:11.600985] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:09:38.293 [2024-12-07 05:37:11.601003] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:38.293 [2024-12-07 05:37:11.601278] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:09:38.293 [2024-12-07 05:37:11.601419] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:09:38.293 [2024-12-07 05:37:11.601432] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:09:38.293 [2024-12-07 05:37:11.601557] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.293 "name": "raid_bdev1", 00:09:38.293 "uuid": "8060b988-04c3-4685-a7a2-43d8ee3d541d", 00:09:38.293 "strip_size_kb": 64, 00:09:38.293 "state": "online", 00:09:38.293 "raid_level": "raid0", 00:09:38.293 "superblock": true, 00:09:38.293 "num_base_bdevs": 4, 00:09:38.293 "num_base_bdevs_discovered": 4, 00:09:38.293 "num_base_bdevs_operational": 4, 00:09:38.293 "base_bdevs_list": [ 00:09:38.293 { 00:09:38.293 "name": "BaseBdev1", 00:09:38.293 "uuid": "905c4f44-5369-581e-abd6-68bd811905c8", 00:09:38.293 "is_configured": true, 00:09:38.293 "data_offset": 2048, 00:09:38.293 "data_size": 63488 00:09:38.293 }, 00:09:38.293 { 00:09:38.293 "name": "BaseBdev2", 00:09:38.293 "uuid": "f0ba9352-1bf3-5957-88d9-fd7136866d44", 00:09:38.293 "is_configured": true, 00:09:38.293 "data_offset": 2048, 00:09:38.293 "data_size": 63488 00:09:38.293 }, 00:09:38.293 { 00:09:38.293 "name": "BaseBdev3", 00:09:38.293 "uuid": "2797385c-820d-55c0-bfde-1e2c6a920099", 00:09:38.293 "is_configured": true, 00:09:38.293 "data_offset": 2048, 00:09:38.293 "data_size": 63488 00:09:38.293 }, 00:09:38.293 { 00:09:38.293 "name": "BaseBdev4", 00:09:38.293 "uuid": "c8e9d75c-0440-5d76-80e0-e64915f4d25e", 00:09:38.293 "is_configured": true, 00:09:38.293 "data_offset": 2048, 00:09:38.293 "data_size": 63488 00:09:38.293 } 00:09:38.293 ] 00:09:38.293 }' 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.293 05:37:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.862 05:37:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:38.862 05:37:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:38.862 [2024-12-07 05:37:12.102363] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.795 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:39.795 "name": "raid_bdev1", 00:09:39.795 "uuid": "8060b988-04c3-4685-a7a2-43d8ee3d541d", 00:09:39.795 "strip_size_kb": 64, 00:09:39.795 "state": "online", 00:09:39.795 "raid_level": "raid0", 00:09:39.795 "superblock": true, 00:09:39.795 "num_base_bdevs": 4, 00:09:39.795 "num_base_bdevs_discovered": 4, 00:09:39.795 "num_base_bdevs_operational": 4, 00:09:39.795 "base_bdevs_list": [ 00:09:39.795 { 00:09:39.795 "name": "BaseBdev1", 00:09:39.795 "uuid": "905c4f44-5369-581e-abd6-68bd811905c8", 00:09:39.795 "is_configured": true, 00:09:39.795 "data_offset": 2048, 00:09:39.795 "data_size": 63488 00:09:39.795 }, 00:09:39.795 { 00:09:39.795 "name": "BaseBdev2", 00:09:39.795 "uuid": "f0ba9352-1bf3-5957-88d9-fd7136866d44", 00:09:39.795 "is_configured": true, 00:09:39.795 "data_offset": 2048, 00:09:39.795 "data_size": 63488 00:09:39.795 }, 00:09:39.795 { 00:09:39.795 "name": "BaseBdev3", 00:09:39.796 "uuid": "2797385c-820d-55c0-bfde-1e2c6a920099", 00:09:39.796 "is_configured": true, 00:09:39.796 "data_offset": 2048, 00:09:39.796 "data_size": 63488 00:09:39.796 }, 00:09:39.796 { 00:09:39.796 "name": "BaseBdev4", 00:09:39.796 "uuid": "c8e9d75c-0440-5d76-80e0-e64915f4d25e", 00:09:39.796 "is_configured": true, 00:09:39.796 "data_offset": 2048, 00:09:39.796 "data_size": 63488 00:09:39.796 } 00:09:39.796 ] 00:09:39.796 }' 00:09:39.796 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:39.796 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.362 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:40.362 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.362 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.362 [2024-12-07 05:37:13.470884] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:40.362 [2024-12-07 05:37:13.470969] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:40.362 [2024-12-07 05:37:13.473790] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:40.362 [2024-12-07 05:37:13.473847] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:40.362 [2024-12-07 05:37:13.473898] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:40.362 [2024-12-07 05:37:13.473909] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:09:40.362 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.362 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81559 00:09:40.362 { 00:09:40.362 "results": [ 00:09:40.362 { 00:09:40.362 "job": "raid_bdev1", 00:09:40.362 "core_mask": "0x1", 00:09:40.363 "workload": "randrw", 00:09:40.363 "percentage": 50, 00:09:40.363 "status": "finished", 00:09:40.363 "queue_depth": 1, 00:09:40.363 "io_size": 131072, 00:09:40.363 "runtime": 1.369084, 00:09:40.363 "iops": 15370.130685918468, 00:09:40.363 "mibps": 1921.2663357398085, 00:09:40.363 "io_failed": 1, 00:09:40.363 "io_timeout": 0, 00:09:40.363 "avg_latency_us": 89.907771531306, 00:09:40.363 "min_latency_us": 27.276855895196505, 00:09:40.363 "max_latency_us": 1609.7816593886462 00:09:40.363 } 00:09:40.363 ], 00:09:40.363 "core_count": 1 00:09:40.363 } 00:09:40.363 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 81559 ']' 00:09:40.363 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 81559 00:09:40.363 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:40.363 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:40.363 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81559 00:09:40.363 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:40.363 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:40.363 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81559' 00:09:40.363 killing process with pid 81559 00:09:40.363 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 81559 00:09:40.363 [2024-12-07 05:37:13.517083] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:40.363 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 81559 00:09:40.363 [2024-12-07 05:37:13.552146] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:40.651 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.ltCeWtklGa 00:09:40.651 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:40.651 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:40.651 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:09:40.651 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:40.651 ************************************ 00:09:40.651 END TEST raid_read_error_test 00:09:40.651 ************************************ 00:09:40.651 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:40.651 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:40.651 05:37:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:09:40.651 00:09:40.651 real 0m3.256s 00:09:40.651 user 0m4.085s 00:09:40.651 sys 0m0.520s 00:09:40.651 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:40.651 05:37:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.651 05:37:13 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:09:40.651 05:37:13 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:40.651 05:37:13 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:40.651 05:37:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:40.651 ************************************ 00:09:40.651 START TEST raid_write_error_test 00:09:40.651 ************************************ 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 write 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.PVu9dTP8Mb 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81688 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81688 00:09:40.651 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 81688 ']' 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:40.651 05:37:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.651 [2024-12-07 05:37:13.936620] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:09:40.651 [2024-12-07 05:37:13.936765] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81688 ] 00:09:40.963 [2024-12-07 05:37:14.090298] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:40.963 [2024-12-07 05:37:14.116819] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:40.963 [2024-12-07 05:37:14.159167] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:40.963 [2024-12-07 05:37:14.159303] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.534 BaseBdev1_malloc 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.534 true 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.534 [2024-12-07 05:37:14.806692] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:41.534 [2024-12-07 05:37:14.806795] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.534 [2024-12-07 05:37:14.806854] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:41.534 [2024-12-07 05:37:14.806884] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.534 [2024-12-07 05:37:14.809115] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.534 [2024-12-07 05:37:14.809188] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:41.534 BaseBdev1 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.534 BaseBdev2_malloc 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.534 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.535 true 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.535 [2024-12-07 05:37:14.847317] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:41.535 [2024-12-07 05:37:14.847372] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.535 [2024-12-07 05:37:14.847406] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:41.535 [2024-12-07 05:37:14.847423] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.535 [2024-12-07 05:37:14.849601] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.535 [2024-12-07 05:37:14.849648] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:41.535 BaseBdev2 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.535 BaseBdev3_malloc 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.535 true 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.535 [2024-12-07 05:37:14.887859] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:41.535 [2024-12-07 05:37:14.887910] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.535 [2024-12-07 05:37:14.887929] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:41.535 [2024-12-07 05:37:14.887938] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.535 [2024-12-07 05:37:14.890044] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.535 [2024-12-07 05:37:14.890078] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:41.535 BaseBdev3 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.535 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.795 BaseBdev4_malloc 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.795 true 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.795 [2024-12-07 05:37:14.936186] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:09:41.795 [2024-12-07 05:37:14.936238] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.795 [2024-12-07 05:37:14.936260] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:41.795 [2024-12-07 05:37:14.936269] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.795 [2024-12-07 05:37:14.938354] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.795 [2024-12-07 05:37:14.938389] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:09:41.795 BaseBdev4 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.795 [2024-12-07 05:37:14.948208] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:41.795 [2024-12-07 05:37:14.950106] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:41.795 [2024-12-07 05:37:14.950195] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:41.795 [2024-12-07 05:37:14.950246] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:41.795 [2024-12-07 05:37:14.950433] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:09:41.795 [2024-12-07 05:37:14.950444] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:41.795 [2024-12-07 05:37:14.950751] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:09:41.795 [2024-12-07 05:37:14.950896] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:09:41.795 [2024-12-07 05:37:14.950915] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:09:41.795 [2024-12-07 05:37:14.951050] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:41.795 05:37:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.795 05:37:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.795 "name": "raid_bdev1", 00:09:41.795 "uuid": "2a292f24-a3e8-43a9-bf0e-8e11e067e5b4", 00:09:41.795 "strip_size_kb": 64, 00:09:41.795 "state": "online", 00:09:41.795 "raid_level": "raid0", 00:09:41.795 "superblock": true, 00:09:41.795 "num_base_bdevs": 4, 00:09:41.795 "num_base_bdevs_discovered": 4, 00:09:41.795 "num_base_bdevs_operational": 4, 00:09:41.795 "base_bdevs_list": [ 00:09:41.795 { 00:09:41.795 "name": "BaseBdev1", 00:09:41.795 "uuid": "7d7bb8da-b6fd-55b5-a420-9cf7a20c530c", 00:09:41.795 "is_configured": true, 00:09:41.795 "data_offset": 2048, 00:09:41.795 "data_size": 63488 00:09:41.795 }, 00:09:41.795 { 00:09:41.795 "name": "BaseBdev2", 00:09:41.795 "uuid": "aa48432b-9b24-5bde-a640-aab815291053", 00:09:41.795 "is_configured": true, 00:09:41.795 "data_offset": 2048, 00:09:41.795 "data_size": 63488 00:09:41.795 }, 00:09:41.795 { 00:09:41.795 "name": "BaseBdev3", 00:09:41.795 "uuid": "ef3a1d76-d482-5771-9930-b21f9788fafe", 00:09:41.795 "is_configured": true, 00:09:41.795 "data_offset": 2048, 00:09:41.795 "data_size": 63488 00:09:41.795 }, 00:09:41.795 { 00:09:41.795 "name": "BaseBdev4", 00:09:41.795 "uuid": "3e594274-18af-53b3-ac4a-a5fa49f376b2", 00:09:41.795 "is_configured": true, 00:09:41.795 "data_offset": 2048, 00:09:41.795 "data_size": 63488 00:09:41.795 } 00:09:41.795 ] 00:09:41.795 }' 00:09:41.795 05:37:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.795 05:37:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.054 05:37:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:42.054 05:37:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:42.312 [2024-12-07 05:37:15.511645] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:43.253 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.254 "name": "raid_bdev1", 00:09:43.254 "uuid": "2a292f24-a3e8-43a9-bf0e-8e11e067e5b4", 00:09:43.254 "strip_size_kb": 64, 00:09:43.254 "state": "online", 00:09:43.254 "raid_level": "raid0", 00:09:43.254 "superblock": true, 00:09:43.254 "num_base_bdevs": 4, 00:09:43.254 "num_base_bdevs_discovered": 4, 00:09:43.254 "num_base_bdevs_operational": 4, 00:09:43.254 "base_bdevs_list": [ 00:09:43.254 { 00:09:43.254 "name": "BaseBdev1", 00:09:43.254 "uuid": "7d7bb8da-b6fd-55b5-a420-9cf7a20c530c", 00:09:43.254 "is_configured": true, 00:09:43.254 "data_offset": 2048, 00:09:43.254 "data_size": 63488 00:09:43.254 }, 00:09:43.254 { 00:09:43.254 "name": "BaseBdev2", 00:09:43.254 "uuid": "aa48432b-9b24-5bde-a640-aab815291053", 00:09:43.254 "is_configured": true, 00:09:43.254 "data_offset": 2048, 00:09:43.254 "data_size": 63488 00:09:43.254 }, 00:09:43.254 { 00:09:43.254 "name": "BaseBdev3", 00:09:43.254 "uuid": "ef3a1d76-d482-5771-9930-b21f9788fafe", 00:09:43.254 "is_configured": true, 00:09:43.254 "data_offset": 2048, 00:09:43.254 "data_size": 63488 00:09:43.254 }, 00:09:43.254 { 00:09:43.254 "name": "BaseBdev4", 00:09:43.254 "uuid": "3e594274-18af-53b3-ac4a-a5fa49f376b2", 00:09:43.254 "is_configured": true, 00:09:43.254 "data_offset": 2048, 00:09:43.254 "data_size": 63488 00:09:43.254 } 00:09:43.254 ] 00:09:43.254 }' 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.254 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.517 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:43.517 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.517 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.517 [2024-12-07 05:37:16.835922] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:43.517 [2024-12-07 05:37:16.836020] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:43.517 [2024-12-07 05:37:16.838729] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:43.517 [2024-12-07 05:37:16.838778] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:43.517 [2024-12-07 05:37:16.838844] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:43.517 [2024-12-07 05:37:16.838855] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:09:43.517 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.517 { 00:09:43.517 "results": [ 00:09:43.517 { 00:09:43.517 "job": "raid_bdev1", 00:09:43.517 "core_mask": "0x1", 00:09:43.517 "workload": "randrw", 00:09:43.517 "percentage": 50, 00:09:43.517 "status": "finished", 00:09:43.517 "queue_depth": 1, 00:09:43.517 "io_size": 131072, 00:09:43.517 "runtime": 1.325106, 00:09:43.517 "iops": 15572.339118530896, 00:09:43.517 "mibps": 1946.542389816362, 00:09:43.517 "io_failed": 1, 00:09:43.517 "io_timeout": 0, 00:09:43.517 "avg_latency_us": 88.77281555699075, 00:09:43.517 "min_latency_us": 26.829694323144103, 00:09:43.517 "max_latency_us": 1667.0183406113538 00:09:43.517 } 00:09:43.517 ], 00:09:43.517 "core_count": 1 00:09:43.517 } 00:09:43.517 05:37:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81688 00:09:43.517 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 81688 ']' 00:09:43.517 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 81688 00:09:43.517 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:43.517 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:43.517 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81688 00:09:43.776 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:43.776 killing process with pid 81688 00:09:43.776 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:43.776 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81688' 00:09:43.776 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 81688 00:09:43.776 [2024-12-07 05:37:16.886312] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:43.776 05:37:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 81688 00:09:43.776 [2024-12-07 05:37:16.922582] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:43.776 05:37:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:43.776 05:37:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.PVu9dTP8Mb 00:09:43.776 05:37:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:43.776 05:37:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.75 00:09:43.776 05:37:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:43.776 05:37:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:43.776 05:37:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:43.776 ************************************ 00:09:43.776 END TEST raid_write_error_test 00:09:43.776 ************************************ 00:09:43.776 05:37:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.75 != \0\.\0\0 ]] 00:09:43.776 00:09:43.776 real 0m3.305s 00:09:43.776 user 0m4.185s 00:09:43.776 sys 0m0.517s 00:09:43.776 05:37:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:43.776 05:37:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.035 05:37:17 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:44.035 05:37:17 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:09:44.035 05:37:17 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:44.035 05:37:17 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:44.035 05:37:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:44.035 ************************************ 00:09:44.035 START TEST raid_state_function_test 00:09:44.035 ************************************ 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 false 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=81821 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:44.035 Process raid pid: 81821 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81821' 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 81821 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 81821 ']' 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:44.035 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:44.035 05:37:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.035 [2024-12-07 05:37:17.298209] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:09:44.035 [2024-12-07 05:37:17.298336] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:44.294 [2024-12-07 05:37:17.432092] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:44.294 [2024-12-07 05:37:17.459141] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:44.294 [2024-12-07 05:37:17.501244] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:44.294 [2024-12-07 05:37:17.501281] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.862 [2024-12-07 05:37:18.155554] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:44.862 [2024-12-07 05:37:18.155620] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:44.862 [2024-12-07 05:37:18.155641] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:44.862 [2024-12-07 05:37:18.155652] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:44.862 [2024-12-07 05:37:18.155658] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:44.862 [2024-12-07 05:37:18.155668] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:44.862 [2024-12-07 05:37:18.155674] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:44.862 [2024-12-07 05:37:18.155682] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:44.862 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:44.863 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:44.863 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:44.863 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:44.863 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.863 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:44.863 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.863 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.863 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.863 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:44.863 "name": "Existed_Raid", 00:09:44.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:44.863 "strip_size_kb": 64, 00:09:44.863 "state": "configuring", 00:09:44.863 "raid_level": "concat", 00:09:44.863 "superblock": false, 00:09:44.863 "num_base_bdevs": 4, 00:09:44.863 "num_base_bdevs_discovered": 0, 00:09:44.863 "num_base_bdevs_operational": 4, 00:09:44.863 "base_bdevs_list": [ 00:09:44.863 { 00:09:44.863 "name": "BaseBdev1", 00:09:44.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:44.863 "is_configured": false, 00:09:44.863 "data_offset": 0, 00:09:44.863 "data_size": 0 00:09:44.863 }, 00:09:44.863 { 00:09:44.863 "name": "BaseBdev2", 00:09:44.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:44.863 "is_configured": false, 00:09:44.863 "data_offset": 0, 00:09:44.863 "data_size": 0 00:09:44.863 }, 00:09:44.863 { 00:09:44.863 "name": "BaseBdev3", 00:09:44.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:44.863 "is_configured": false, 00:09:44.863 "data_offset": 0, 00:09:44.863 "data_size": 0 00:09:44.863 }, 00:09:44.863 { 00:09:44.863 "name": "BaseBdev4", 00:09:44.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:44.863 "is_configured": false, 00:09:44.863 "data_offset": 0, 00:09:44.863 "data_size": 0 00:09:44.863 } 00:09:44.863 ] 00:09:44.863 }' 00:09:44.863 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:44.863 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.432 [2024-12-07 05:37:18.650688] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:45.432 [2024-12-07 05:37:18.650789] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.432 [2024-12-07 05:37:18.662677] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:45.432 [2024-12-07 05:37:18.662760] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:45.432 [2024-12-07 05:37:18.662810] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:45.432 [2024-12-07 05:37:18.662860] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:45.432 [2024-12-07 05:37:18.662886] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:45.432 [2024-12-07 05:37:18.662909] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:45.432 [2024-12-07 05:37:18.662950] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:45.432 [2024-12-07 05:37:18.662984] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.432 [2024-12-07 05:37:18.683854] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:45.432 BaseBdev1 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.432 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.433 [ 00:09:45.433 { 00:09:45.433 "name": "BaseBdev1", 00:09:45.433 "aliases": [ 00:09:45.433 "0b5907de-e420-47b5-a070-a082baaf9ac0" 00:09:45.433 ], 00:09:45.433 "product_name": "Malloc disk", 00:09:45.433 "block_size": 512, 00:09:45.433 "num_blocks": 65536, 00:09:45.433 "uuid": "0b5907de-e420-47b5-a070-a082baaf9ac0", 00:09:45.433 "assigned_rate_limits": { 00:09:45.433 "rw_ios_per_sec": 0, 00:09:45.433 "rw_mbytes_per_sec": 0, 00:09:45.433 "r_mbytes_per_sec": 0, 00:09:45.433 "w_mbytes_per_sec": 0 00:09:45.433 }, 00:09:45.433 "claimed": true, 00:09:45.433 "claim_type": "exclusive_write", 00:09:45.433 "zoned": false, 00:09:45.433 "supported_io_types": { 00:09:45.433 "read": true, 00:09:45.433 "write": true, 00:09:45.433 "unmap": true, 00:09:45.433 "flush": true, 00:09:45.433 "reset": true, 00:09:45.433 "nvme_admin": false, 00:09:45.433 "nvme_io": false, 00:09:45.433 "nvme_io_md": false, 00:09:45.433 "write_zeroes": true, 00:09:45.433 "zcopy": true, 00:09:45.433 "get_zone_info": false, 00:09:45.433 "zone_management": false, 00:09:45.433 "zone_append": false, 00:09:45.433 "compare": false, 00:09:45.433 "compare_and_write": false, 00:09:45.433 "abort": true, 00:09:45.433 "seek_hole": false, 00:09:45.433 "seek_data": false, 00:09:45.433 "copy": true, 00:09:45.433 "nvme_iov_md": false 00:09:45.433 }, 00:09:45.433 "memory_domains": [ 00:09:45.433 { 00:09:45.433 "dma_device_id": "system", 00:09:45.433 "dma_device_type": 1 00:09:45.433 }, 00:09:45.433 { 00:09:45.433 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:45.433 "dma_device_type": 2 00:09:45.433 } 00:09:45.433 ], 00:09:45.433 "driver_specific": {} 00:09:45.433 } 00:09:45.433 ] 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.433 "name": "Existed_Raid", 00:09:45.433 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.433 "strip_size_kb": 64, 00:09:45.433 "state": "configuring", 00:09:45.433 "raid_level": "concat", 00:09:45.433 "superblock": false, 00:09:45.433 "num_base_bdevs": 4, 00:09:45.433 "num_base_bdevs_discovered": 1, 00:09:45.433 "num_base_bdevs_operational": 4, 00:09:45.433 "base_bdevs_list": [ 00:09:45.433 { 00:09:45.433 "name": "BaseBdev1", 00:09:45.433 "uuid": "0b5907de-e420-47b5-a070-a082baaf9ac0", 00:09:45.433 "is_configured": true, 00:09:45.433 "data_offset": 0, 00:09:45.433 "data_size": 65536 00:09:45.433 }, 00:09:45.433 { 00:09:45.433 "name": "BaseBdev2", 00:09:45.433 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.433 "is_configured": false, 00:09:45.433 "data_offset": 0, 00:09:45.433 "data_size": 0 00:09:45.433 }, 00:09:45.433 { 00:09:45.433 "name": "BaseBdev3", 00:09:45.433 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.433 "is_configured": false, 00:09:45.433 "data_offset": 0, 00:09:45.433 "data_size": 0 00:09:45.433 }, 00:09:45.433 { 00:09:45.433 "name": "BaseBdev4", 00:09:45.433 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.433 "is_configured": false, 00:09:45.433 "data_offset": 0, 00:09:45.433 "data_size": 0 00:09:45.433 } 00:09:45.433 ] 00:09:45.433 }' 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.433 05:37:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.003 [2024-12-07 05:37:19.159108] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:46.003 [2024-12-07 05:37:19.159161] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.003 [2024-12-07 05:37:19.167128] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:46.003 [2024-12-07 05:37:19.169027] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:46.003 [2024-12-07 05:37:19.169115] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:46.003 [2024-12-07 05:37:19.169129] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:46.003 [2024-12-07 05:37:19.169139] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:46.003 [2024-12-07 05:37:19.169146] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:46.003 [2024-12-07 05:37:19.169154] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:46.003 "name": "Existed_Raid", 00:09:46.003 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.003 "strip_size_kb": 64, 00:09:46.003 "state": "configuring", 00:09:46.003 "raid_level": "concat", 00:09:46.003 "superblock": false, 00:09:46.003 "num_base_bdevs": 4, 00:09:46.003 "num_base_bdevs_discovered": 1, 00:09:46.003 "num_base_bdevs_operational": 4, 00:09:46.003 "base_bdevs_list": [ 00:09:46.003 { 00:09:46.003 "name": "BaseBdev1", 00:09:46.003 "uuid": "0b5907de-e420-47b5-a070-a082baaf9ac0", 00:09:46.003 "is_configured": true, 00:09:46.003 "data_offset": 0, 00:09:46.003 "data_size": 65536 00:09:46.003 }, 00:09:46.003 { 00:09:46.003 "name": "BaseBdev2", 00:09:46.003 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.003 "is_configured": false, 00:09:46.003 "data_offset": 0, 00:09:46.003 "data_size": 0 00:09:46.003 }, 00:09:46.003 { 00:09:46.003 "name": "BaseBdev3", 00:09:46.003 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.003 "is_configured": false, 00:09:46.003 "data_offset": 0, 00:09:46.003 "data_size": 0 00:09:46.003 }, 00:09:46.003 { 00:09:46.003 "name": "BaseBdev4", 00:09:46.003 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.003 "is_configured": false, 00:09:46.003 "data_offset": 0, 00:09:46.003 "data_size": 0 00:09:46.003 } 00:09:46.003 ] 00:09:46.003 }' 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:46.003 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.262 [2024-12-07 05:37:19.545327] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:46.262 BaseBdev2 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.262 [ 00:09:46.262 { 00:09:46.262 "name": "BaseBdev2", 00:09:46.262 "aliases": [ 00:09:46.262 "2277e25f-cb57-4e38-8419-b869c567a5f6" 00:09:46.262 ], 00:09:46.262 "product_name": "Malloc disk", 00:09:46.262 "block_size": 512, 00:09:46.262 "num_blocks": 65536, 00:09:46.262 "uuid": "2277e25f-cb57-4e38-8419-b869c567a5f6", 00:09:46.262 "assigned_rate_limits": { 00:09:46.262 "rw_ios_per_sec": 0, 00:09:46.262 "rw_mbytes_per_sec": 0, 00:09:46.262 "r_mbytes_per_sec": 0, 00:09:46.262 "w_mbytes_per_sec": 0 00:09:46.262 }, 00:09:46.262 "claimed": true, 00:09:46.262 "claim_type": "exclusive_write", 00:09:46.262 "zoned": false, 00:09:46.262 "supported_io_types": { 00:09:46.262 "read": true, 00:09:46.262 "write": true, 00:09:46.262 "unmap": true, 00:09:46.262 "flush": true, 00:09:46.262 "reset": true, 00:09:46.262 "nvme_admin": false, 00:09:46.262 "nvme_io": false, 00:09:46.262 "nvme_io_md": false, 00:09:46.262 "write_zeroes": true, 00:09:46.262 "zcopy": true, 00:09:46.262 "get_zone_info": false, 00:09:46.262 "zone_management": false, 00:09:46.262 "zone_append": false, 00:09:46.262 "compare": false, 00:09:46.262 "compare_and_write": false, 00:09:46.262 "abort": true, 00:09:46.262 "seek_hole": false, 00:09:46.262 "seek_data": false, 00:09:46.262 "copy": true, 00:09:46.262 "nvme_iov_md": false 00:09:46.262 }, 00:09:46.262 "memory_domains": [ 00:09:46.262 { 00:09:46.262 "dma_device_id": "system", 00:09:46.262 "dma_device_type": 1 00:09:46.262 }, 00:09:46.262 { 00:09:46.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.262 "dma_device_type": 2 00:09:46.262 } 00:09:46.262 ], 00:09:46.262 "driver_specific": {} 00:09:46.262 } 00:09:46.262 ] 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.262 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.263 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.523 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:46.523 "name": "Existed_Raid", 00:09:46.523 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.523 "strip_size_kb": 64, 00:09:46.523 "state": "configuring", 00:09:46.523 "raid_level": "concat", 00:09:46.523 "superblock": false, 00:09:46.523 "num_base_bdevs": 4, 00:09:46.523 "num_base_bdevs_discovered": 2, 00:09:46.523 "num_base_bdevs_operational": 4, 00:09:46.523 "base_bdevs_list": [ 00:09:46.523 { 00:09:46.523 "name": "BaseBdev1", 00:09:46.523 "uuid": "0b5907de-e420-47b5-a070-a082baaf9ac0", 00:09:46.523 "is_configured": true, 00:09:46.523 "data_offset": 0, 00:09:46.523 "data_size": 65536 00:09:46.523 }, 00:09:46.523 { 00:09:46.523 "name": "BaseBdev2", 00:09:46.523 "uuid": "2277e25f-cb57-4e38-8419-b869c567a5f6", 00:09:46.523 "is_configured": true, 00:09:46.523 "data_offset": 0, 00:09:46.523 "data_size": 65536 00:09:46.523 }, 00:09:46.523 { 00:09:46.523 "name": "BaseBdev3", 00:09:46.523 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.523 "is_configured": false, 00:09:46.523 "data_offset": 0, 00:09:46.523 "data_size": 0 00:09:46.523 }, 00:09:46.523 { 00:09:46.523 "name": "BaseBdev4", 00:09:46.523 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.523 "is_configured": false, 00:09:46.523 "data_offset": 0, 00:09:46.523 "data_size": 0 00:09:46.523 } 00:09:46.523 ] 00:09:46.523 }' 00:09:46.523 05:37:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:46.523 05:37:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.783 [2024-12-07 05:37:20.037008] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:46.783 BaseBdev3 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.783 [ 00:09:46.783 { 00:09:46.783 "name": "BaseBdev3", 00:09:46.783 "aliases": [ 00:09:46.783 "05a7e934-d2ff-4669-9714-14b103547065" 00:09:46.783 ], 00:09:46.783 "product_name": "Malloc disk", 00:09:46.783 "block_size": 512, 00:09:46.783 "num_blocks": 65536, 00:09:46.783 "uuid": "05a7e934-d2ff-4669-9714-14b103547065", 00:09:46.783 "assigned_rate_limits": { 00:09:46.783 "rw_ios_per_sec": 0, 00:09:46.783 "rw_mbytes_per_sec": 0, 00:09:46.783 "r_mbytes_per_sec": 0, 00:09:46.783 "w_mbytes_per_sec": 0 00:09:46.783 }, 00:09:46.783 "claimed": true, 00:09:46.783 "claim_type": "exclusive_write", 00:09:46.783 "zoned": false, 00:09:46.783 "supported_io_types": { 00:09:46.783 "read": true, 00:09:46.783 "write": true, 00:09:46.783 "unmap": true, 00:09:46.783 "flush": true, 00:09:46.783 "reset": true, 00:09:46.783 "nvme_admin": false, 00:09:46.783 "nvme_io": false, 00:09:46.783 "nvme_io_md": false, 00:09:46.783 "write_zeroes": true, 00:09:46.783 "zcopy": true, 00:09:46.783 "get_zone_info": false, 00:09:46.783 "zone_management": false, 00:09:46.783 "zone_append": false, 00:09:46.783 "compare": false, 00:09:46.783 "compare_and_write": false, 00:09:46.783 "abort": true, 00:09:46.783 "seek_hole": false, 00:09:46.783 "seek_data": false, 00:09:46.783 "copy": true, 00:09:46.783 "nvme_iov_md": false 00:09:46.783 }, 00:09:46.783 "memory_domains": [ 00:09:46.783 { 00:09:46.783 "dma_device_id": "system", 00:09:46.783 "dma_device_type": 1 00:09:46.783 }, 00:09:46.783 { 00:09:46.783 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.783 "dma_device_type": 2 00:09:46.783 } 00:09:46.783 ], 00:09:46.783 "driver_specific": {} 00:09:46.783 } 00:09:46.783 ] 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.783 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:46.783 "name": "Existed_Raid", 00:09:46.783 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.783 "strip_size_kb": 64, 00:09:46.783 "state": "configuring", 00:09:46.783 "raid_level": "concat", 00:09:46.783 "superblock": false, 00:09:46.783 "num_base_bdevs": 4, 00:09:46.783 "num_base_bdevs_discovered": 3, 00:09:46.783 "num_base_bdevs_operational": 4, 00:09:46.783 "base_bdevs_list": [ 00:09:46.783 { 00:09:46.783 "name": "BaseBdev1", 00:09:46.783 "uuid": "0b5907de-e420-47b5-a070-a082baaf9ac0", 00:09:46.783 "is_configured": true, 00:09:46.783 "data_offset": 0, 00:09:46.783 "data_size": 65536 00:09:46.784 }, 00:09:46.784 { 00:09:46.784 "name": "BaseBdev2", 00:09:46.784 "uuid": "2277e25f-cb57-4e38-8419-b869c567a5f6", 00:09:46.784 "is_configured": true, 00:09:46.784 "data_offset": 0, 00:09:46.784 "data_size": 65536 00:09:46.784 }, 00:09:46.784 { 00:09:46.784 "name": "BaseBdev3", 00:09:46.784 "uuid": "05a7e934-d2ff-4669-9714-14b103547065", 00:09:46.784 "is_configured": true, 00:09:46.784 "data_offset": 0, 00:09:46.784 "data_size": 65536 00:09:46.784 }, 00:09:46.784 { 00:09:46.784 "name": "BaseBdev4", 00:09:46.784 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.784 "is_configured": false, 00:09:46.784 "data_offset": 0, 00:09:46.784 "data_size": 0 00:09:46.784 } 00:09:46.784 ] 00:09:46.784 }' 00:09:46.784 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:46.784 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.353 [2024-12-07 05:37:20.523145] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:47.353 [2024-12-07 05:37:20.523280] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:47.353 [2024-12-07 05:37:20.523293] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:47.353 [2024-12-07 05:37:20.523601] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:47.353 [2024-12-07 05:37:20.523763] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:47.353 [2024-12-07 05:37:20.523777] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:47.353 [2024-12-07 05:37:20.524003] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:47.353 BaseBdev4 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.353 [ 00:09:47.353 { 00:09:47.353 "name": "BaseBdev4", 00:09:47.353 "aliases": [ 00:09:47.353 "4e65bb6e-0e04-4146-b729-b1aba04b831b" 00:09:47.353 ], 00:09:47.353 "product_name": "Malloc disk", 00:09:47.353 "block_size": 512, 00:09:47.353 "num_blocks": 65536, 00:09:47.353 "uuid": "4e65bb6e-0e04-4146-b729-b1aba04b831b", 00:09:47.353 "assigned_rate_limits": { 00:09:47.353 "rw_ios_per_sec": 0, 00:09:47.353 "rw_mbytes_per_sec": 0, 00:09:47.353 "r_mbytes_per_sec": 0, 00:09:47.353 "w_mbytes_per_sec": 0 00:09:47.353 }, 00:09:47.353 "claimed": true, 00:09:47.353 "claim_type": "exclusive_write", 00:09:47.353 "zoned": false, 00:09:47.353 "supported_io_types": { 00:09:47.353 "read": true, 00:09:47.353 "write": true, 00:09:47.353 "unmap": true, 00:09:47.353 "flush": true, 00:09:47.353 "reset": true, 00:09:47.353 "nvme_admin": false, 00:09:47.353 "nvme_io": false, 00:09:47.353 "nvme_io_md": false, 00:09:47.353 "write_zeroes": true, 00:09:47.353 "zcopy": true, 00:09:47.353 "get_zone_info": false, 00:09:47.353 "zone_management": false, 00:09:47.353 "zone_append": false, 00:09:47.353 "compare": false, 00:09:47.353 "compare_and_write": false, 00:09:47.353 "abort": true, 00:09:47.353 "seek_hole": false, 00:09:47.353 "seek_data": false, 00:09:47.353 "copy": true, 00:09:47.353 "nvme_iov_md": false 00:09:47.353 }, 00:09:47.353 "memory_domains": [ 00:09:47.353 { 00:09:47.353 "dma_device_id": "system", 00:09:47.353 "dma_device_type": 1 00:09:47.353 }, 00:09:47.353 { 00:09:47.353 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.353 "dma_device_type": 2 00:09:47.353 } 00:09:47.353 ], 00:09:47.353 "driver_specific": {} 00:09:47.353 } 00:09:47.353 ] 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.353 "name": "Existed_Raid", 00:09:47.353 "uuid": "52f3c0ae-439a-437e-99f6-ee621b9f6c61", 00:09:47.353 "strip_size_kb": 64, 00:09:47.353 "state": "online", 00:09:47.353 "raid_level": "concat", 00:09:47.353 "superblock": false, 00:09:47.353 "num_base_bdevs": 4, 00:09:47.353 "num_base_bdevs_discovered": 4, 00:09:47.353 "num_base_bdevs_operational": 4, 00:09:47.353 "base_bdevs_list": [ 00:09:47.353 { 00:09:47.353 "name": "BaseBdev1", 00:09:47.353 "uuid": "0b5907de-e420-47b5-a070-a082baaf9ac0", 00:09:47.353 "is_configured": true, 00:09:47.353 "data_offset": 0, 00:09:47.353 "data_size": 65536 00:09:47.353 }, 00:09:47.353 { 00:09:47.353 "name": "BaseBdev2", 00:09:47.353 "uuid": "2277e25f-cb57-4e38-8419-b869c567a5f6", 00:09:47.353 "is_configured": true, 00:09:47.353 "data_offset": 0, 00:09:47.353 "data_size": 65536 00:09:47.353 }, 00:09:47.353 { 00:09:47.353 "name": "BaseBdev3", 00:09:47.353 "uuid": "05a7e934-d2ff-4669-9714-14b103547065", 00:09:47.353 "is_configured": true, 00:09:47.353 "data_offset": 0, 00:09:47.353 "data_size": 65536 00:09:47.353 }, 00:09:47.353 { 00:09:47.353 "name": "BaseBdev4", 00:09:47.353 "uuid": "4e65bb6e-0e04-4146-b729-b1aba04b831b", 00:09:47.353 "is_configured": true, 00:09:47.353 "data_offset": 0, 00:09:47.353 "data_size": 65536 00:09:47.353 } 00:09:47.353 ] 00:09:47.353 }' 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.353 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.923 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:47.923 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:47.923 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:47.923 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:47.923 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:47.923 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:47.923 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:47.923 05:37:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:47.923 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.923 05:37:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.923 [2024-12-07 05:37:20.998741] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:47.923 "name": "Existed_Raid", 00:09:47.923 "aliases": [ 00:09:47.923 "52f3c0ae-439a-437e-99f6-ee621b9f6c61" 00:09:47.923 ], 00:09:47.923 "product_name": "Raid Volume", 00:09:47.923 "block_size": 512, 00:09:47.923 "num_blocks": 262144, 00:09:47.923 "uuid": "52f3c0ae-439a-437e-99f6-ee621b9f6c61", 00:09:47.923 "assigned_rate_limits": { 00:09:47.923 "rw_ios_per_sec": 0, 00:09:47.923 "rw_mbytes_per_sec": 0, 00:09:47.923 "r_mbytes_per_sec": 0, 00:09:47.923 "w_mbytes_per_sec": 0 00:09:47.923 }, 00:09:47.923 "claimed": false, 00:09:47.923 "zoned": false, 00:09:47.923 "supported_io_types": { 00:09:47.923 "read": true, 00:09:47.923 "write": true, 00:09:47.923 "unmap": true, 00:09:47.923 "flush": true, 00:09:47.923 "reset": true, 00:09:47.923 "nvme_admin": false, 00:09:47.923 "nvme_io": false, 00:09:47.923 "nvme_io_md": false, 00:09:47.923 "write_zeroes": true, 00:09:47.923 "zcopy": false, 00:09:47.923 "get_zone_info": false, 00:09:47.923 "zone_management": false, 00:09:47.923 "zone_append": false, 00:09:47.923 "compare": false, 00:09:47.923 "compare_and_write": false, 00:09:47.923 "abort": false, 00:09:47.923 "seek_hole": false, 00:09:47.923 "seek_data": false, 00:09:47.923 "copy": false, 00:09:47.923 "nvme_iov_md": false 00:09:47.923 }, 00:09:47.923 "memory_domains": [ 00:09:47.923 { 00:09:47.923 "dma_device_id": "system", 00:09:47.923 "dma_device_type": 1 00:09:47.923 }, 00:09:47.923 { 00:09:47.923 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.923 "dma_device_type": 2 00:09:47.923 }, 00:09:47.923 { 00:09:47.923 "dma_device_id": "system", 00:09:47.923 "dma_device_type": 1 00:09:47.923 }, 00:09:47.923 { 00:09:47.923 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.923 "dma_device_type": 2 00:09:47.923 }, 00:09:47.923 { 00:09:47.923 "dma_device_id": "system", 00:09:47.923 "dma_device_type": 1 00:09:47.923 }, 00:09:47.923 { 00:09:47.923 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.923 "dma_device_type": 2 00:09:47.923 }, 00:09:47.923 { 00:09:47.923 "dma_device_id": "system", 00:09:47.923 "dma_device_type": 1 00:09:47.923 }, 00:09:47.923 { 00:09:47.923 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.923 "dma_device_type": 2 00:09:47.923 } 00:09:47.923 ], 00:09:47.923 "driver_specific": { 00:09:47.923 "raid": { 00:09:47.923 "uuid": "52f3c0ae-439a-437e-99f6-ee621b9f6c61", 00:09:47.923 "strip_size_kb": 64, 00:09:47.923 "state": "online", 00:09:47.923 "raid_level": "concat", 00:09:47.923 "superblock": false, 00:09:47.923 "num_base_bdevs": 4, 00:09:47.923 "num_base_bdevs_discovered": 4, 00:09:47.923 "num_base_bdevs_operational": 4, 00:09:47.923 "base_bdevs_list": [ 00:09:47.923 { 00:09:47.923 "name": "BaseBdev1", 00:09:47.923 "uuid": "0b5907de-e420-47b5-a070-a082baaf9ac0", 00:09:47.923 "is_configured": true, 00:09:47.923 "data_offset": 0, 00:09:47.923 "data_size": 65536 00:09:47.923 }, 00:09:47.923 { 00:09:47.923 "name": "BaseBdev2", 00:09:47.923 "uuid": "2277e25f-cb57-4e38-8419-b869c567a5f6", 00:09:47.923 "is_configured": true, 00:09:47.923 "data_offset": 0, 00:09:47.923 "data_size": 65536 00:09:47.923 }, 00:09:47.923 { 00:09:47.923 "name": "BaseBdev3", 00:09:47.923 "uuid": "05a7e934-d2ff-4669-9714-14b103547065", 00:09:47.923 "is_configured": true, 00:09:47.923 "data_offset": 0, 00:09:47.923 "data_size": 65536 00:09:47.923 }, 00:09:47.923 { 00:09:47.923 "name": "BaseBdev4", 00:09:47.923 "uuid": "4e65bb6e-0e04-4146-b729-b1aba04b831b", 00:09:47.923 "is_configured": true, 00:09:47.923 "data_offset": 0, 00:09:47.923 "data_size": 65536 00:09:47.923 } 00:09:47.923 ] 00:09:47.923 } 00:09:47.923 } 00:09:47.923 }' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:47.923 BaseBdev2 00:09:47.923 BaseBdev3 00:09:47.923 BaseBdev4' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.923 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.183 [2024-12-07 05:37:21.317940] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:48.183 [2024-12-07 05:37:21.318018] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:48.183 [2024-12-07 05:37:21.318134] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.183 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.183 "name": "Existed_Raid", 00:09:48.183 "uuid": "52f3c0ae-439a-437e-99f6-ee621b9f6c61", 00:09:48.183 "strip_size_kb": 64, 00:09:48.183 "state": "offline", 00:09:48.183 "raid_level": "concat", 00:09:48.183 "superblock": false, 00:09:48.183 "num_base_bdevs": 4, 00:09:48.183 "num_base_bdevs_discovered": 3, 00:09:48.183 "num_base_bdevs_operational": 3, 00:09:48.183 "base_bdevs_list": [ 00:09:48.183 { 00:09:48.183 "name": null, 00:09:48.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.183 "is_configured": false, 00:09:48.183 "data_offset": 0, 00:09:48.183 "data_size": 65536 00:09:48.183 }, 00:09:48.183 { 00:09:48.183 "name": "BaseBdev2", 00:09:48.183 "uuid": "2277e25f-cb57-4e38-8419-b869c567a5f6", 00:09:48.183 "is_configured": true, 00:09:48.183 "data_offset": 0, 00:09:48.183 "data_size": 65536 00:09:48.183 }, 00:09:48.183 { 00:09:48.183 "name": "BaseBdev3", 00:09:48.183 "uuid": "05a7e934-d2ff-4669-9714-14b103547065", 00:09:48.183 "is_configured": true, 00:09:48.183 "data_offset": 0, 00:09:48.183 "data_size": 65536 00:09:48.183 }, 00:09:48.183 { 00:09:48.183 "name": "BaseBdev4", 00:09:48.183 "uuid": "4e65bb6e-0e04-4146-b729-b1aba04b831b", 00:09:48.183 "is_configured": true, 00:09:48.183 "data_offset": 0, 00:09:48.183 "data_size": 65536 00:09:48.183 } 00:09:48.183 ] 00:09:48.183 }' 00:09:48.184 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.184 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.443 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:48.443 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:48.443 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.443 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.443 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.443 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.703 [2024-12-07 05:37:21.852724] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.703 [2024-12-07 05:37:21.919873] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.703 [2024-12-07 05:37:21.982971] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:48.703 [2024-12-07 05:37:21.983016] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.703 05:37:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.703 BaseBdev2 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.703 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.964 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.964 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:48.964 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.964 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.964 [ 00:09:48.964 { 00:09:48.964 "name": "BaseBdev2", 00:09:48.964 "aliases": [ 00:09:48.964 "a0fde126-b014-4f66-bf31-72c4c1681481" 00:09:48.964 ], 00:09:48.964 "product_name": "Malloc disk", 00:09:48.964 "block_size": 512, 00:09:48.964 "num_blocks": 65536, 00:09:48.964 "uuid": "a0fde126-b014-4f66-bf31-72c4c1681481", 00:09:48.964 "assigned_rate_limits": { 00:09:48.964 "rw_ios_per_sec": 0, 00:09:48.964 "rw_mbytes_per_sec": 0, 00:09:48.964 "r_mbytes_per_sec": 0, 00:09:48.964 "w_mbytes_per_sec": 0 00:09:48.964 }, 00:09:48.964 "claimed": false, 00:09:48.964 "zoned": false, 00:09:48.964 "supported_io_types": { 00:09:48.964 "read": true, 00:09:48.964 "write": true, 00:09:48.964 "unmap": true, 00:09:48.964 "flush": true, 00:09:48.964 "reset": true, 00:09:48.964 "nvme_admin": false, 00:09:48.964 "nvme_io": false, 00:09:48.964 "nvme_io_md": false, 00:09:48.964 "write_zeroes": true, 00:09:48.964 "zcopy": true, 00:09:48.964 "get_zone_info": false, 00:09:48.964 "zone_management": false, 00:09:48.964 "zone_append": false, 00:09:48.964 "compare": false, 00:09:48.964 "compare_and_write": false, 00:09:48.964 "abort": true, 00:09:48.964 "seek_hole": false, 00:09:48.964 "seek_data": false, 00:09:48.964 "copy": true, 00:09:48.964 "nvme_iov_md": false 00:09:48.964 }, 00:09:48.964 "memory_domains": [ 00:09:48.964 { 00:09:48.965 "dma_device_id": "system", 00:09:48.965 "dma_device_type": 1 00:09:48.965 }, 00:09:48.965 { 00:09:48.965 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.965 "dma_device_type": 2 00:09:48.965 } 00:09:48.965 ], 00:09:48.965 "driver_specific": {} 00:09:48.965 } 00:09:48.965 ] 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.965 BaseBdev3 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.965 [ 00:09:48.965 { 00:09:48.965 "name": "BaseBdev3", 00:09:48.965 "aliases": [ 00:09:48.965 "cabc74c9-3490-472f-8243-a6ef5cd0dc7d" 00:09:48.965 ], 00:09:48.965 "product_name": "Malloc disk", 00:09:48.965 "block_size": 512, 00:09:48.965 "num_blocks": 65536, 00:09:48.965 "uuid": "cabc74c9-3490-472f-8243-a6ef5cd0dc7d", 00:09:48.965 "assigned_rate_limits": { 00:09:48.965 "rw_ios_per_sec": 0, 00:09:48.965 "rw_mbytes_per_sec": 0, 00:09:48.965 "r_mbytes_per_sec": 0, 00:09:48.965 "w_mbytes_per_sec": 0 00:09:48.965 }, 00:09:48.965 "claimed": false, 00:09:48.965 "zoned": false, 00:09:48.965 "supported_io_types": { 00:09:48.965 "read": true, 00:09:48.965 "write": true, 00:09:48.965 "unmap": true, 00:09:48.965 "flush": true, 00:09:48.965 "reset": true, 00:09:48.965 "nvme_admin": false, 00:09:48.965 "nvme_io": false, 00:09:48.965 "nvme_io_md": false, 00:09:48.965 "write_zeroes": true, 00:09:48.965 "zcopy": true, 00:09:48.965 "get_zone_info": false, 00:09:48.965 "zone_management": false, 00:09:48.965 "zone_append": false, 00:09:48.965 "compare": false, 00:09:48.965 "compare_and_write": false, 00:09:48.965 "abort": true, 00:09:48.965 "seek_hole": false, 00:09:48.965 "seek_data": false, 00:09:48.965 "copy": true, 00:09:48.965 "nvme_iov_md": false 00:09:48.965 }, 00:09:48.965 "memory_domains": [ 00:09:48.965 { 00:09:48.965 "dma_device_id": "system", 00:09:48.965 "dma_device_type": 1 00:09:48.965 }, 00:09:48.965 { 00:09:48.965 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.965 "dma_device_type": 2 00:09:48.965 } 00:09:48.965 ], 00:09:48.965 "driver_specific": {} 00:09:48.965 } 00:09:48.965 ] 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.965 BaseBdev4 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.965 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.965 [ 00:09:48.965 { 00:09:48.965 "name": "BaseBdev4", 00:09:48.965 "aliases": [ 00:09:48.965 "5c474af9-d235-4584-8a46-102030535195" 00:09:48.965 ], 00:09:48.965 "product_name": "Malloc disk", 00:09:48.965 "block_size": 512, 00:09:48.965 "num_blocks": 65536, 00:09:48.965 "uuid": "5c474af9-d235-4584-8a46-102030535195", 00:09:48.965 "assigned_rate_limits": { 00:09:48.965 "rw_ios_per_sec": 0, 00:09:48.965 "rw_mbytes_per_sec": 0, 00:09:48.965 "r_mbytes_per_sec": 0, 00:09:48.965 "w_mbytes_per_sec": 0 00:09:48.965 }, 00:09:48.965 "claimed": false, 00:09:48.965 "zoned": false, 00:09:48.965 "supported_io_types": { 00:09:48.965 "read": true, 00:09:48.965 "write": true, 00:09:48.965 "unmap": true, 00:09:48.965 "flush": true, 00:09:48.965 "reset": true, 00:09:48.965 "nvme_admin": false, 00:09:48.965 "nvme_io": false, 00:09:48.965 "nvme_io_md": false, 00:09:48.965 "write_zeroes": true, 00:09:48.965 "zcopy": true, 00:09:48.965 "get_zone_info": false, 00:09:48.965 "zone_management": false, 00:09:48.966 "zone_append": false, 00:09:48.966 "compare": false, 00:09:48.966 "compare_and_write": false, 00:09:48.966 "abort": true, 00:09:48.966 "seek_hole": false, 00:09:48.966 "seek_data": false, 00:09:48.966 "copy": true, 00:09:48.966 "nvme_iov_md": false 00:09:48.966 }, 00:09:48.966 "memory_domains": [ 00:09:48.966 { 00:09:48.966 "dma_device_id": "system", 00:09:48.966 "dma_device_type": 1 00:09:48.966 }, 00:09:48.966 { 00:09:48.966 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.966 "dma_device_type": 2 00:09:48.966 } 00:09:48.966 ], 00:09:48.966 "driver_specific": {} 00:09:48.966 } 00:09:48.966 ] 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.966 [2024-12-07 05:37:22.210819] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:48.966 [2024-12-07 05:37:22.210908] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:48.966 [2024-12-07 05:37:22.210966] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:48.966 [2024-12-07 05:37:22.212828] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:48.966 [2024-12-07 05:37:22.212876] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.966 "name": "Existed_Raid", 00:09:48.966 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.966 "strip_size_kb": 64, 00:09:48.966 "state": "configuring", 00:09:48.966 "raid_level": "concat", 00:09:48.966 "superblock": false, 00:09:48.966 "num_base_bdevs": 4, 00:09:48.966 "num_base_bdevs_discovered": 3, 00:09:48.966 "num_base_bdevs_operational": 4, 00:09:48.966 "base_bdevs_list": [ 00:09:48.966 { 00:09:48.966 "name": "BaseBdev1", 00:09:48.966 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.966 "is_configured": false, 00:09:48.966 "data_offset": 0, 00:09:48.966 "data_size": 0 00:09:48.966 }, 00:09:48.966 { 00:09:48.966 "name": "BaseBdev2", 00:09:48.966 "uuid": "a0fde126-b014-4f66-bf31-72c4c1681481", 00:09:48.966 "is_configured": true, 00:09:48.966 "data_offset": 0, 00:09:48.966 "data_size": 65536 00:09:48.966 }, 00:09:48.966 { 00:09:48.966 "name": "BaseBdev3", 00:09:48.966 "uuid": "cabc74c9-3490-472f-8243-a6ef5cd0dc7d", 00:09:48.966 "is_configured": true, 00:09:48.966 "data_offset": 0, 00:09:48.966 "data_size": 65536 00:09:48.966 }, 00:09:48.966 { 00:09:48.966 "name": "BaseBdev4", 00:09:48.966 "uuid": "5c474af9-d235-4584-8a46-102030535195", 00:09:48.966 "is_configured": true, 00:09:48.966 "data_offset": 0, 00:09:48.966 "data_size": 65536 00:09:48.966 } 00:09:48.966 ] 00:09:48.966 }' 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.966 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.535 [2024-12-07 05:37:22.622204] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.535 "name": "Existed_Raid", 00:09:49.535 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.535 "strip_size_kb": 64, 00:09:49.535 "state": "configuring", 00:09:49.535 "raid_level": "concat", 00:09:49.535 "superblock": false, 00:09:49.535 "num_base_bdevs": 4, 00:09:49.535 "num_base_bdevs_discovered": 2, 00:09:49.535 "num_base_bdevs_operational": 4, 00:09:49.535 "base_bdevs_list": [ 00:09:49.535 { 00:09:49.535 "name": "BaseBdev1", 00:09:49.535 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.535 "is_configured": false, 00:09:49.535 "data_offset": 0, 00:09:49.535 "data_size": 0 00:09:49.535 }, 00:09:49.535 { 00:09:49.535 "name": null, 00:09:49.535 "uuid": "a0fde126-b014-4f66-bf31-72c4c1681481", 00:09:49.535 "is_configured": false, 00:09:49.535 "data_offset": 0, 00:09:49.535 "data_size": 65536 00:09:49.535 }, 00:09:49.535 { 00:09:49.535 "name": "BaseBdev3", 00:09:49.535 "uuid": "cabc74c9-3490-472f-8243-a6ef5cd0dc7d", 00:09:49.535 "is_configured": true, 00:09:49.535 "data_offset": 0, 00:09:49.535 "data_size": 65536 00:09:49.535 }, 00:09:49.535 { 00:09:49.535 "name": "BaseBdev4", 00:09:49.535 "uuid": "5c474af9-d235-4584-8a46-102030535195", 00:09:49.535 "is_configured": true, 00:09:49.535 "data_offset": 0, 00:09:49.535 "data_size": 65536 00:09:49.535 } 00:09:49.535 ] 00:09:49.535 }' 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.535 05:37:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.795 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.795 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.795 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.795 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:49.795 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.795 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:49.795 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:49.795 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.795 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.795 [2024-12-07 05:37:23.084375] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:49.795 BaseBdev1 00:09:49.795 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.796 [ 00:09:49.796 { 00:09:49.796 "name": "BaseBdev1", 00:09:49.796 "aliases": [ 00:09:49.796 "6c649e4c-bd38-4f57-bdbc-6c915807083c" 00:09:49.796 ], 00:09:49.796 "product_name": "Malloc disk", 00:09:49.796 "block_size": 512, 00:09:49.796 "num_blocks": 65536, 00:09:49.796 "uuid": "6c649e4c-bd38-4f57-bdbc-6c915807083c", 00:09:49.796 "assigned_rate_limits": { 00:09:49.796 "rw_ios_per_sec": 0, 00:09:49.796 "rw_mbytes_per_sec": 0, 00:09:49.796 "r_mbytes_per_sec": 0, 00:09:49.796 "w_mbytes_per_sec": 0 00:09:49.796 }, 00:09:49.796 "claimed": true, 00:09:49.796 "claim_type": "exclusive_write", 00:09:49.796 "zoned": false, 00:09:49.796 "supported_io_types": { 00:09:49.796 "read": true, 00:09:49.796 "write": true, 00:09:49.796 "unmap": true, 00:09:49.796 "flush": true, 00:09:49.796 "reset": true, 00:09:49.796 "nvme_admin": false, 00:09:49.796 "nvme_io": false, 00:09:49.796 "nvme_io_md": false, 00:09:49.796 "write_zeroes": true, 00:09:49.796 "zcopy": true, 00:09:49.796 "get_zone_info": false, 00:09:49.796 "zone_management": false, 00:09:49.796 "zone_append": false, 00:09:49.796 "compare": false, 00:09:49.796 "compare_and_write": false, 00:09:49.796 "abort": true, 00:09:49.796 "seek_hole": false, 00:09:49.796 "seek_data": false, 00:09:49.796 "copy": true, 00:09:49.796 "nvme_iov_md": false 00:09:49.796 }, 00:09:49.796 "memory_domains": [ 00:09:49.796 { 00:09:49.796 "dma_device_id": "system", 00:09:49.796 "dma_device_type": 1 00:09:49.796 }, 00:09:49.796 { 00:09:49.796 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.796 "dma_device_type": 2 00:09:49.796 } 00:09:49.796 ], 00:09:49.796 "driver_specific": {} 00:09:49.796 } 00:09:49.796 ] 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.796 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.054 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.054 "name": "Existed_Raid", 00:09:50.054 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.054 "strip_size_kb": 64, 00:09:50.054 "state": "configuring", 00:09:50.054 "raid_level": "concat", 00:09:50.054 "superblock": false, 00:09:50.054 "num_base_bdevs": 4, 00:09:50.054 "num_base_bdevs_discovered": 3, 00:09:50.054 "num_base_bdevs_operational": 4, 00:09:50.054 "base_bdevs_list": [ 00:09:50.054 { 00:09:50.054 "name": "BaseBdev1", 00:09:50.054 "uuid": "6c649e4c-bd38-4f57-bdbc-6c915807083c", 00:09:50.054 "is_configured": true, 00:09:50.054 "data_offset": 0, 00:09:50.054 "data_size": 65536 00:09:50.054 }, 00:09:50.054 { 00:09:50.054 "name": null, 00:09:50.054 "uuid": "a0fde126-b014-4f66-bf31-72c4c1681481", 00:09:50.054 "is_configured": false, 00:09:50.054 "data_offset": 0, 00:09:50.054 "data_size": 65536 00:09:50.054 }, 00:09:50.054 { 00:09:50.054 "name": "BaseBdev3", 00:09:50.054 "uuid": "cabc74c9-3490-472f-8243-a6ef5cd0dc7d", 00:09:50.054 "is_configured": true, 00:09:50.054 "data_offset": 0, 00:09:50.054 "data_size": 65536 00:09:50.054 }, 00:09:50.054 { 00:09:50.054 "name": "BaseBdev4", 00:09:50.054 "uuid": "5c474af9-d235-4584-8a46-102030535195", 00:09:50.054 "is_configured": true, 00:09:50.054 "data_offset": 0, 00:09:50.054 "data_size": 65536 00:09:50.054 } 00:09:50.054 ] 00:09:50.054 }' 00:09:50.054 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.054 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.313 [2024-12-07 05:37:23.623519] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.313 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.313 "name": "Existed_Raid", 00:09:50.313 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.313 "strip_size_kb": 64, 00:09:50.313 "state": "configuring", 00:09:50.313 "raid_level": "concat", 00:09:50.313 "superblock": false, 00:09:50.313 "num_base_bdevs": 4, 00:09:50.313 "num_base_bdevs_discovered": 2, 00:09:50.313 "num_base_bdevs_operational": 4, 00:09:50.313 "base_bdevs_list": [ 00:09:50.313 { 00:09:50.313 "name": "BaseBdev1", 00:09:50.313 "uuid": "6c649e4c-bd38-4f57-bdbc-6c915807083c", 00:09:50.313 "is_configured": true, 00:09:50.313 "data_offset": 0, 00:09:50.313 "data_size": 65536 00:09:50.313 }, 00:09:50.313 { 00:09:50.313 "name": null, 00:09:50.313 "uuid": "a0fde126-b014-4f66-bf31-72c4c1681481", 00:09:50.313 "is_configured": false, 00:09:50.313 "data_offset": 0, 00:09:50.313 "data_size": 65536 00:09:50.313 }, 00:09:50.313 { 00:09:50.313 "name": null, 00:09:50.313 "uuid": "cabc74c9-3490-472f-8243-a6ef5cd0dc7d", 00:09:50.313 "is_configured": false, 00:09:50.313 "data_offset": 0, 00:09:50.314 "data_size": 65536 00:09:50.314 }, 00:09:50.314 { 00:09:50.314 "name": "BaseBdev4", 00:09:50.314 "uuid": "5c474af9-d235-4584-8a46-102030535195", 00:09:50.314 "is_configured": true, 00:09:50.314 "data_offset": 0, 00:09:50.314 "data_size": 65536 00:09:50.314 } 00:09:50.314 ] 00:09:50.314 }' 00:09:50.314 05:37:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.314 05:37:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.881 [2024-12-07 05:37:24.102745] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.881 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.881 "name": "Existed_Raid", 00:09:50.881 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.881 "strip_size_kb": 64, 00:09:50.881 "state": "configuring", 00:09:50.882 "raid_level": "concat", 00:09:50.882 "superblock": false, 00:09:50.882 "num_base_bdevs": 4, 00:09:50.882 "num_base_bdevs_discovered": 3, 00:09:50.882 "num_base_bdevs_operational": 4, 00:09:50.882 "base_bdevs_list": [ 00:09:50.882 { 00:09:50.882 "name": "BaseBdev1", 00:09:50.882 "uuid": "6c649e4c-bd38-4f57-bdbc-6c915807083c", 00:09:50.882 "is_configured": true, 00:09:50.882 "data_offset": 0, 00:09:50.882 "data_size": 65536 00:09:50.882 }, 00:09:50.882 { 00:09:50.882 "name": null, 00:09:50.882 "uuid": "a0fde126-b014-4f66-bf31-72c4c1681481", 00:09:50.882 "is_configured": false, 00:09:50.882 "data_offset": 0, 00:09:50.882 "data_size": 65536 00:09:50.882 }, 00:09:50.882 { 00:09:50.882 "name": "BaseBdev3", 00:09:50.882 "uuid": "cabc74c9-3490-472f-8243-a6ef5cd0dc7d", 00:09:50.882 "is_configured": true, 00:09:50.882 "data_offset": 0, 00:09:50.882 "data_size": 65536 00:09:50.882 }, 00:09:50.882 { 00:09:50.882 "name": "BaseBdev4", 00:09:50.882 "uuid": "5c474af9-d235-4584-8a46-102030535195", 00:09:50.882 "is_configured": true, 00:09:50.882 "data_offset": 0, 00:09:50.882 "data_size": 65536 00:09:50.882 } 00:09:50.882 ] 00:09:50.882 }' 00:09:50.882 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.882 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.140 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.140 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.140 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.140 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.399 [2024-12-07 05:37:24.554010] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.399 "name": "Existed_Raid", 00:09:51.399 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.399 "strip_size_kb": 64, 00:09:51.399 "state": "configuring", 00:09:51.399 "raid_level": "concat", 00:09:51.399 "superblock": false, 00:09:51.399 "num_base_bdevs": 4, 00:09:51.399 "num_base_bdevs_discovered": 2, 00:09:51.399 "num_base_bdevs_operational": 4, 00:09:51.399 "base_bdevs_list": [ 00:09:51.399 { 00:09:51.399 "name": null, 00:09:51.399 "uuid": "6c649e4c-bd38-4f57-bdbc-6c915807083c", 00:09:51.399 "is_configured": false, 00:09:51.399 "data_offset": 0, 00:09:51.399 "data_size": 65536 00:09:51.399 }, 00:09:51.399 { 00:09:51.399 "name": null, 00:09:51.399 "uuid": "a0fde126-b014-4f66-bf31-72c4c1681481", 00:09:51.399 "is_configured": false, 00:09:51.399 "data_offset": 0, 00:09:51.399 "data_size": 65536 00:09:51.399 }, 00:09:51.399 { 00:09:51.399 "name": "BaseBdev3", 00:09:51.399 "uuid": "cabc74c9-3490-472f-8243-a6ef5cd0dc7d", 00:09:51.399 "is_configured": true, 00:09:51.399 "data_offset": 0, 00:09:51.399 "data_size": 65536 00:09:51.399 }, 00:09:51.399 { 00:09:51.399 "name": "BaseBdev4", 00:09:51.399 "uuid": "5c474af9-d235-4584-8a46-102030535195", 00:09:51.399 "is_configured": true, 00:09:51.399 "data_offset": 0, 00:09:51.399 "data_size": 65536 00:09:51.399 } 00:09:51.399 ] 00:09:51.399 }' 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.399 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.658 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.658 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.658 05:37:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.658 05:37:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:51.658 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.925 [2024-12-07 05:37:25.043768] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.925 "name": "Existed_Raid", 00:09:51.925 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.925 "strip_size_kb": 64, 00:09:51.925 "state": "configuring", 00:09:51.925 "raid_level": "concat", 00:09:51.925 "superblock": false, 00:09:51.925 "num_base_bdevs": 4, 00:09:51.925 "num_base_bdevs_discovered": 3, 00:09:51.925 "num_base_bdevs_operational": 4, 00:09:51.925 "base_bdevs_list": [ 00:09:51.925 { 00:09:51.925 "name": null, 00:09:51.925 "uuid": "6c649e4c-bd38-4f57-bdbc-6c915807083c", 00:09:51.925 "is_configured": false, 00:09:51.925 "data_offset": 0, 00:09:51.925 "data_size": 65536 00:09:51.925 }, 00:09:51.925 { 00:09:51.925 "name": "BaseBdev2", 00:09:51.925 "uuid": "a0fde126-b014-4f66-bf31-72c4c1681481", 00:09:51.925 "is_configured": true, 00:09:51.925 "data_offset": 0, 00:09:51.925 "data_size": 65536 00:09:51.925 }, 00:09:51.925 { 00:09:51.925 "name": "BaseBdev3", 00:09:51.925 "uuid": "cabc74c9-3490-472f-8243-a6ef5cd0dc7d", 00:09:51.925 "is_configured": true, 00:09:51.925 "data_offset": 0, 00:09:51.925 "data_size": 65536 00:09:51.925 }, 00:09:51.925 { 00:09:51.925 "name": "BaseBdev4", 00:09:51.925 "uuid": "5c474af9-d235-4584-8a46-102030535195", 00:09:51.925 "is_configured": true, 00:09:51.925 "data_offset": 0, 00:09:51.925 "data_size": 65536 00:09:51.925 } 00:09:51.925 ] 00:09:51.925 }' 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.925 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 6c649e4c-bd38-4f57-bdbc-6c915807083c 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.190 [2024-12-07 05:37:25.509965] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:52.190 [2024-12-07 05:37:25.510010] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:52.190 [2024-12-07 05:37:25.510018] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:52.190 [2024-12-07 05:37:25.510291] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:52.190 [2024-12-07 05:37:25.510403] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:52.190 [2024-12-07 05:37:25.510415] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:52.190 [2024-12-07 05:37:25.510609] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:52.190 NewBaseBdev 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.190 [ 00:09:52.190 { 00:09:52.190 "name": "NewBaseBdev", 00:09:52.190 "aliases": [ 00:09:52.190 "6c649e4c-bd38-4f57-bdbc-6c915807083c" 00:09:52.190 ], 00:09:52.190 "product_name": "Malloc disk", 00:09:52.190 "block_size": 512, 00:09:52.190 "num_blocks": 65536, 00:09:52.190 "uuid": "6c649e4c-bd38-4f57-bdbc-6c915807083c", 00:09:52.190 "assigned_rate_limits": { 00:09:52.190 "rw_ios_per_sec": 0, 00:09:52.190 "rw_mbytes_per_sec": 0, 00:09:52.190 "r_mbytes_per_sec": 0, 00:09:52.190 "w_mbytes_per_sec": 0 00:09:52.190 }, 00:09:52.190 "claimed": true, 00:09:52.190 "claim_type": "exclusive_write", 00:09:52.190 "zoned": false, 00:09:52.190 "supported_io_types": { 00:09:52.190 "read": true, 00:09:52.190 "write": true, 00:09:52.190 "unmap": true, 00:09:52.190 "flush": true, 00:09:52.190 "reset": true, 00:09:52.190 "nvme_admin": false, 00:09:52.190 "nvme_io": false, 00:09:52.190 "nvme_io_md": false, 00:09:52.190 "write_zeroes": true, 00:09:52.190 "zcopy": true, 00:09:52.190 "get_zone_info": false, 00:09:52.190 "zone_management": false, 00:09:52.190 "zone_append": false, 00:09:52.190 "compare": false, 00:09:52.190 "compare_and_write": false, 00:09:52.190 "abort": true, 00:09:52.190 "seek_hole": false, 00:09:52.190 "seek_data": false, 00:09:52.190 "copy": true, 00:09:52.190 "nvme_iov_md": false 00:09:52.190 }, 00:09:52.190 "memory_domains": [ 00:09:52.190 { 00:09:52.190 "dma_device_id": "system", 00:09:52.190 "dma_device_type": 1 00:09:52.190 }, 00:09:52.190 { 00:09:52.190 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.190 "dma_device_type": 2 00:09:52.190 } 00:09:52.190 ], 00:09:52.190 "driver_specific": {} 00:09:52.190 } 00:09:52.190 ] 00:09:52.190 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.454 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.454 "name": "Existed_Raid", 00:09:52.454 "uuid": "9832e661-6b6f-433c-b38a-d3465c0648af", 00:09:52.454 "strip_size_kb": 64, 00:09:52.454 "state": "online", 00:09:52.454 "raid_level": "concat", 00:09:52.454 "superblock": false, 00:09:52.454 "num_base_bdevs": 4, 00:09:52.454 "num_base_bdevs_discovered": 4, 00:09:52.454 "num_base_bdevs_operational": 4, 00:09:52.454 "base_bdevs_list": [ 00:09:52.454 { 00:09:52.454 "name": "NewBaseBdev", 00:09:52.454 "uuid": "6c649e4c-bd38-4f57-bdbc-6c915807083c", 00:09:52.454 "is_configured": true, 00:09:52.454 "data_offset": 0, 00:09:52.454 "data_size": 65536 00:09:52.454 }, 00:09:52.454 { 00:09:52.454 "name": "BaseBdev2", 00:09:52.454 "uuid": "a0fde126-b014-4f66-bf31-72c4c1681481", 00:09:52.454 "is_configured": true, 00:09:52.454 "data_offset": 0, 00:09:52.454 "data_size": 65536 00:09:52.454 }, 00:09:52.454 { 00:09:52.454 "name": "BaseBdev3", 00:09:52.454 "uuid": "cabc74c9-3490-472f-8243-a6ef5cd0dc7d", 00:09:52.455 "is_configured": true, 00:09:52.455 "data_offset": 0, 00:09:52.455 "data_size": 65536 00:09:52.455 }, 00:09:52.455 { 00:09:52.455 "name": "BaseBdev4", 00:09:52.455 "uuid": "5c474af9-d235-4584-8a46-102030535195", 00:09:52.455 "is_configured": true, 00:09:52.455 "data_offset": 0, 00:09:52.455 "data_size": 65536 00:09:52.455 } 00:09:52.455 ] 00:09:52.455 }' 00:09:52.455 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.455 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.713 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:52.713 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:52.713 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:52.713 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:52.713 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:52.713 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:52.713 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:52.713 05:37:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:52.713 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.713 05:37:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.713 [2024-12-07 05:37:25.993529] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:52.713 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.713 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:52.713 "name": "Existed_Raid", 00:09:52.713 "aliases": [ 00:09:52.713 "9832e661-6b6f-433c-b38a-d3465c0648af" 00:09:52.713 ], 00:09:52.713 "product_name": "Raid Volume", 00:09:52.713 "block_size": 512, 00:09:52.713 "num_blocks": 262144, 00:09:52.713 "uuid": "9832e661-6b6f-433c-b38a-d3465c0648af", 00:09:52.713 "assigned_rate_limits": { 00:09:52.713 "rw_ios_per_sec": 0, 00:09:52.713 "rw_mbytes_per_sec": 0, 00:09:52.713 "r_mbytes_per_sec": 0, 00:09:52.713 "w_mbytes_per_sec": 0 00:09:52.713 }, 00:09:52.713 "claimed": false, 00:09:52.713 "zoned": false, 00:09:52.713 "supported_io_types": { 00:09:52.713 "read": true, 00:09:52.713 "write": true, 00:09:52.713 "unmap": true, 00:09:52.713 "flush": true, 00:09:52.713 "reset": true, 00:09:52.713 "nvme_admin": false, 00:09:52.713 "nvme_io": false, 00:09:52.713 "nvme_io_md": false, 00:09:52.713 "write_zeroes": true, 00:09:52.713 "zcopy": false, 00:09:52.713 "get_zone_info": false, 00:09:52.713 "zone_management": false, 00:09:52.713 "zone_append": false, 00:09:52.713 "compare": false, 00:09:52.713 "compare_and_write": false, 00:09:52.713 "abort": false, 00:09:52.713 "seek_hole": false, 00:09:52.713 "seek_data": false, 00:09:52.713 "copy": false, 00:09:52.713 "nvme_iov_md": false 00:09:52.713 }, 00:09:52.713 "memory_domains": [ 00:09:52.713 { 00:09:52.713 "dma_device_id": "system", 00:09:52.714 "dma_device_type": 1 00:09:52.714 }, 00:09:52.714 { 00:09:52.714 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.714 "dma_device_type": 2 00:09:52.714 }, 00:09:52.714 { 00:09:52.714 "dma_device_id": "system", 00:09:52.714 "dma_device_type": 1 00:09:52.714 }, 00:09:52.714 { 00:09:52.714 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.714 "dma_device_type": 2 00:09:52.714 }, 00:09:52.714 { 00:09:52.714 "dma_device_id": "system", 00:09:52.714 "dma_device_type": 1 00:09:52.714 }, 00:09:52.714 { 00:09:52.714 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.714 "dma_device_type": 2 00:09:52.714 }, 00:09:52.714 { 00:09:52.714 "dma_device_id": "system", 00:09:52.714 "dma_device_type": 1 00:09:52.714 }, 00:09:52.714 { 00:09:52.714 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.714 "dma_device_type": 2 00:09:52.714 } 00:09:52.714 ], 00:09:52.714 "driver_specific": { 00:09:52.714 "raid": { 00:09:52.714 "uuid": "9832e661-6b6f-433c-b38a-d3465c0648af", 00:09:52.714 "strip_size_kb": 64, 00:09:52.714 "state": "online", 00:09:52.714 "raid_level": "concat", 00:09:52.714 "superblock": false, 00:09:52.714 "num_base_bdevs": 4, 00:09:52.714 "num_base_bdevs_discovered": 4, 00:09:52.714 "num_base_bdevs_operational": 4, 00:09:52.714 "base_bdevs_list": [ 00:09:52.714 { 00:09:52.714 "name": "NewBaseBdev", 00:09:52.714 "uuid": "6c649e4c-bd38-4f57-bdbc-6c915807083c", 00:09:52.714 "is_configured": true, 00:09:52.714 "data_offset": 0, 00:09:52.714 "data_size": 65536 00:09:52.714 }, 00:09:52.714 { 00:09:52.714 "name": "BaseBdev2", 00:09:52.714 "uuid": "a0fde126-b014-4f66-bf31-72c4c1681481", 00:09:52.714 "is_configured": true, 00:09:52.714 "data_offset": 0, 00:09:52.714 "data_size": 65536 00:09:52.714 }, 00:09:52.714 { 00:09:52.714 "name": "BaseBdev3", 00:09:52.714 "uuid": "cabc74c9-3490-472f-8243-a6ef5cd0dc7d", 00:09:52.714 "is_configured": true, 00:09:52.714 "data_offset": 0, 00:09:52.714 "data_size": 65536 00:09:52.714 }, 00:09:52.714 { 00:09:52.714 "name": "BaseBdev4", 00:09:52.714 "uuid": "5c474af9-d235-4584-8a46-102030535195", 00:09:52.714 "is_configured": true, 00:09:52.714 "data_offset": 0, 00:09:52.714 "data_size": 65536 00:09:52.714 } 00:09:52.714 ] 00:09:52.714 } 00:09:52.714 } 00:09:52.714 }' 00:09:52.714 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:52.714 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:52.714 BaseBdev2 00:09:52.714 BaseBdev3 00:09:52.714 BaseBdev4' 00:09:52.714 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.974 [2024-12-07 05:37:26.296699] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:52.974 [2024-12-07 05:37:26.296771] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:52.974 [2024-12-07 05:37:26.296882] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:52.974 [2024-12-07 05:37:26.296953] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:52.974 [2024-12-07 05:37:26.296963] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 81821 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 81821 ']' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 81821 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81821 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81821' 00:09:52.974 killing process with pid 81821 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 81821 00:09:52.974 [2024-12-07 05:37:26.330053] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:52.974 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 81821 00:09:53.233 [2024-12-07 05:37:26.370409] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:53.233 05:37:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:53.233 00:09:53.233 real 0m9.368s 00:09:53.233 user 0m16.083s 00:09:53.233 sys 0m1.841s 00:09:53.233 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:53.233 ************************************ 00:09:53.233 END TEST raid_state_function_test 00:09:53.233 ************************************ 00:09:53.233 05:37:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.492 05:37:26 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:09:53.492 05:37:26 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:53.492 05:37:26 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:53.492 05:37:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:53.492 ************************************ 00:09:53.492 START TEST raid_state_function_test_sb 00:09:53.492 ************************************ 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 true 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=82470 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82470' 00:09:53.492 Process raid pid: 82470 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 82470 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 82470 ']' 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:53.492 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:53.492 05:37:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.492 [2024-12-07 05:37:26.728171] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:09:53.492 [2024-12-07 05:37:26.728376] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:53.762 [2024-12-07 05:37:26.882111] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:53.762 [2024-12-07 05:37:26.907531] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:53.762 [2024-12-07 05:37:26.949737] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:53.762 [2024-12-07 05:37:26.949867] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:54.331 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:54.331 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.332 [2024-12-07 05:37:27.564359] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:54.332 [2024-12-07 05:37:27.564420] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:54.332 [2024-12-07 05:37:27.564430] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:54.332 [2024-12-07 05:37:27.564440] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:54.332 [2024-12-07 05:37:27.564446] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:54.332 [2024-12-07 05:37:27.564460] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:54.332 [2024-12-07 05:37:27.564467] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:54.332 [2024-12-07 05:37:27.564475] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.332 "name": "Existed_Raid", 00:09:54.332 "uuid": "c0db6ed9-fbb7-4135-a3cd-f8381a682bdf", 00:09:54.332 "strip_size_kb": 64, 00:09:54.332 "state": "configuring", 00:09:54.332 "raid_level": "concat", 00:09:54.332 "superblock": true, 00:09:54.332 "num_base_bdevs": 4, 00:09:54.332 "num_base_bdevs_discovered": 0, 00:09:54.332 "num_base_bdevs_operational": 4, 00:09:54.332 "base_bdevs_list": [ 00:09:54.332 { 00:09:54.332 "name": "BaseBdev1", 00:09:54.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.332 "is_configured": false, 00:09:54.332 "data_offset": 0, 00:09:54.332 "data_size": 0 00:09:54.332 }, 00:09:54.332 { 00:09:54.332 "name": "BaseBdev2", 00:09:54.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.332 "is_configured": false, 00:09:54.332 "data_offset": 0, 00:09:54.332 "data_size": 0 00:09:54.332 }, 00:09:54.332 { 00:09:54.332 "name": "BaseBdev3", 00:09:54.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.332 "is_configured": false, 00:09:54.332 "data_offset": 0, 00:09:54.332 "data_size": 0 00:09:54.332 }, 00:09:54.332 { 00:09:54.332 "name": "BaseBdev4", 00:09:54.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.332 "is_configured": false, 00:09:54.332 "data_offset": 0, 00:09:54.332 "data_size": 0 00:09:54.332 } 00:09:54.332 ] 00:09:54.332 }' 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.332 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.901 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:54.901 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.901 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.901 [2024-12-07 05:37:27.975549] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:54.901 [2024-12-07 05:37:27.975685] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:54.901 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.901 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:54.901 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.901 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.901 [2024-12-07 05:37:27.987564] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:54.901 [2024-12-07 05:37:27.987692] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:54.901 [2024-12-07 05:37:27.987726] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:54.901 [2024-12-07 05:37:27.987755] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:54.901 [2024-12-07 05:37:27.987782] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:54.901 [2024-12-07 05:37:27.987827] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:54.901 [2024-12-07 05:37:27.987854] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:54.902 [2024-12-07 05:37:27.987883] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:54.902 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.902 05:37:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:54.902 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.902 05:37:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.902 [2024-12-07 05:37:28.008306] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:54.902 BaseBdev1 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.902 [ 00:09:54.902 { 00:09:54.902 "name": "BaseBdev1", 00:09:54.902 "aliases": [ 00:09:54.902 "ba09c5cb-0355-46d7-82dd-d911c416ca39" 00:09:54.902 ], 00:09:54.902 "product_name": "Malloc disk", 00:09:54.902 "block_size": 512, 00:09:54.902 "num_blocks": 65536, 00:09:54.902 "uuid": "ba09c5cb-0355-46d7-82dd-d911c416ca39", 00:09:54.902 "assigned_rate_limits": { 00:09:54.902 "rw_ios_per_sec": 0, 00:09:54.902 "rw_mbytes_per_sec": 0, 00:09:54.902 "r_mbytes_per_sec": 0, 00:09:54.902 "w_mbytes_per_sec": 0 00:09:54.902 }, 00:09:54.902 "claimed": true, 00:09:54.902 "claim_type": "exclusive_write", 00:09:54.902 "zoned": false, 00:09:54.902 "supported_io_types": { 00:09:54.902 "read": true, 00:09:54.902 "write": true, 00:09:54.902 "unmap": true, 00:09:54.902 "flush": true, 00:09:54.902 "reset": true, 00:09:54.902 "nvme_admin": false, 00:09:54.902 "nvme_io": false, 00:09:54.902 "nvme_io_md": false, 00:09:54.902 "write_zeroes": true, 00:09:54.902 "zcopy": true, 00:09:54.902 "get_zone_info": false, 00:09:54.902 "zone_management": false, 00:09:54.902 "zone_append": false, 00:09:54.902 "compare": false, 00:09:54.902 "compare_and_write": false, 00:09:54.902 "abort": true, 00:09:54.902 "seek_hole": false, 00:09:54.902 "seek_data": false, 00:09:54.902 "copy": true, 00:09:54.902 "nvme_iov_md": false 00:09:54.902 }, 00:09:54.902 "memory_domains": [ 00:09:54.902 { 00:09:54.902 "dma_device_id": "system", 00:09:54.902 "dma_device_type": 1 00:09:54.902 }, 00:09:54.902 { 00:09:54.902 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.902 "dma_device_type": 2 00:09:54.902 } 00:09:54.902 ], 00:09:54.902 "driver_specific": {} 00:09:54.902 } 00:09:54.902 ] 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.902 "name": "Existed_Raid", 00:09:54.902 "uuid": "c6680e95-447c-4f41-8ee9-13859c79b705", 00:09:54.902 "strip_size_kb": 64, 00:09:54.902 "state": "configuring", 00:09:54.902 "raid_level": "concat", 00:09:54.902 "superblock": true, 00:09:54.902 "num_base_bdevs": 4, 00:09:54.902 "num_base_bdevs_discovered": 1, 00:09:54.902 "num_base_bdevs_operational": 4, 00:09:54.902 "base_bdevs_list": [ 00:09:54.902 { 00:09:54.902 "name": "BaseBdev1", 00:09:54.902 "uuid": "ba09c5cb-0355-46d7-82dd-d911c416ca39", 00:09:54.902 "is_configured": true, 00:09:54.902 "data_offset": 2048, 00:09:54.902 "data_size": 63488 00:09:54.902 }, 00:09:54.902 { 00:09:54.902 "name": "BaseBdev2", 00:09:54.902 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.902 "is_configured": false, 00:09:54.902 "data_offset": 0, 00:09:54.902 "data_size": 0 00:09:54.902 }, 00:09:54.902 { 00:09:54.902 "name": "BaseBdev3", 00:09:54.902 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.902 "is_configured": false, 00:09:54.902 "data_offset": 0, 00:09:54.902 "data_size": 0 00:09:54.902 }, 00:09:54.902 { 00:09:54.902 "name": "BaseBdev4", 00:09:54.902 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.902 "is_configured": false, 00:09:54.902 "data_offset": 0, 00:09:54.902 "data_size": 0 00:09:54.902 } 00:09:54.902 ] 00:09:54.902 }' 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.902 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.162 [2024-12-07 05:37:28.475581] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:55.162 [2024-12-07 05:37:28.475735] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.162 [2024-12-07 05:37:28.483612] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:55.162 [2024-12-07 05:37:28.485441] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:55.162 [2024-12-07 05:37:28.485487] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:55.162 [2024-12-07 05:37:28.485498] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:55.162 [2024-12-07 05:37:28.485506] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:55.162 [2024-12-07 05:37:28.485513] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:55.162 [2024-12-07 05:37:28.485521] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.162 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.422 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.422 "name": "Existed_Raid", 00:09:55.422 "uuid": "b720ebac-fe87-416c-8b96-f2271759dd3e", 00:09:55.422 "strip_size_kb": 64, 00:09:55.422 "state": "configuring", 00:09:55.422 "raid_level": "concat", 00:09:55.422 "superblock": true, 00:09:55.422 "num_base_bdevs": 4, 00:09:55.422 "num_base_bdevs_discovered": 1, 00:09:55.422 "num_base_bdevs_operational": 4, 00:09:55.422 "base_bdevs_list": [ 00:09:55.422 { 00:09:55.422 "name": "BaseBdev1", 00:09:55.422 "uuid": "ba09c5cb-0355-46d7-82dd-d911c416ca39", 00:09:55.422 "is_configured": true, 00:09:55.422 "data_offset": 2048, 00:09:55.422 "data_size": 63488 00:09:55.422 }, 00:09:55.422 { 00:09:55.422 "name": "BaseBdev2", 00:09:55.422 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.422 "is_configured": false, 00:09:55.422 "data_offset": 0, 00:09:55.422 "data_size": 0 00:09:55.422 }, 00:09:55.422 { 00:09:55.422 "name": "BaseBdev3", 00:09:55.422 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.422 "is_configured": false, 00:09:55.422 "data_offset": 0, 00:09:55.422 "data_size": 0 00:09:55.422 }, 00:09:55.422 { 00:09:55.422 "name": "BaseBdev4", 00:09:55.422 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.422 "is_configured": false, 00:09:55.422 "data_offset": 0, 00:09:55.422 "data_size": 0 00:09:55.422 } 00:09:55.422 ] 00:09:55.422 }' 00:09:55.422 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.422 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.682 [2024-12-07 05:37:28.933805] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:55.682 BaseBdev2 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.682 [ 00:09:55.682 { 00:09:55.682 "name": "BaseBdev2", 00:09:55.682 "aliases": [ 00:09:55.682 "baa351a9-c8ea-42ee-b7a2-d178b1c7389c" 00:09:55.682 ], 00:09:55.682 "product_name": "Malloc disk", 00:09:55.682 "block_size": 512, 00:09:55.682 "num_blocks": 65536, 00:09:55.682 "uuid": "baa351a9-c8ea-42ee-b7a2-d178b1c7389c", 00:09:55.682 "assigned_rate_limits": { 00:09:55.682 "rw_ios_per_sec": 0, 00:09:55.682 "rw_mbytes_per_sec": 0, 00:09:55.682 "r_mbytes_per_sec": 0, 00:09:55.682 "w_mbytes_per_sec": 0 00:09:55.682 }, 00:09:55.682 "claimed": true, 00:09:55.682 "claim_type": "exclusive_write", 00:09:55.682 "zoned": false, 00:09:55.682 "supported_io_types": { 00:09:55.682 "read": true, 00:09:55.682 "write": true, 00:09:55.682 "unmap": true, 00:09:55.682 "flush": true, 00:09:55.682 "reset": true, 00:09:55.682 "nvme_admin": false, 00:09:55.682 "nvme_io": false, 00:09:55.682 "nvme_io_md": false, 00:09:55.682 "write_zeroes": true, 00:09:55.682 "zcopy": true, 00:09:55.682 "get_zone_info": false, 00:09:55.682 "zone_management": false, 00:09:55.682 "zone_append": false, 00:09:55.682 "compare": false, 00:09:55.682 "compare_and_write": false, 00:09:55.682 "abort": true, 00:09:55.682 "seek_hole": false, 00:09:55.682 "seek_data": false, 00:09:55.682 "copy": true, 00:09:55.682 "nvme_iov_md": false 00:09:55.682 }, 00:09:55.682 "memory_domains": [ 00:09:55.682 { 00:09:55.682 "dma_device_id": "system", 00:09:55.682 "dma_device_type": 1 00:09:55.682 }, 00:09:55.682 { 00:09:55.682 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.682 "dma_device_type": 2 00:09:55.682 } 00:09:55.682 ], 00:09:55.682 "driver_specific": {} 00:09:55.682 } 00:09:55.682 ] 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.682 05:37:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.682 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.682 "name": "Existed_Raid", 00:09:55.682 "uuid": "b720ebac-fe87-416c-8b96-f2271759dd3e", 00:09:55.682 "strip_size_kb": 64, 00:09:55.682 "state": "configuring", 00:09:55.682 "raid_level": "concat", 00:09:55.682 "superblock": true, 00:09:55.682 "num_base_bdevs": 4, 00:09:55.682 "num_base_bdevs_discovered": 2, 00:09:55.682 "num_base_bdevs_operational": 4, 00:09:55.682 "base_bdevs_list": [ 00:09:55.682 { 00:09:55.682 "name": "BaseBdev1", 00:09:55.682 "uuid": "ba09c5cb-0355-46d7-82dd-d911c416ca39", 00:09:55.682 "is_configured": true, 00:09:55.682 "data_offset": 2048, 00:09:55.682 "data_size": 63488 00:09:55.682 }, 00:09:55.682 { 00:09:55.682 "name": "BaseBdev2", 00:09:55.682 "uuid": "baa351a9-c8ea-42ee-b7a2-d178b1c7389c", 00:09:55.683 "is_configured": true, 00:09:55.683 "data_offset": 2048, 00:09:55.683 "data_size": 63488 00:09:55.683 }, 00:09:55.683 { 00:09:55.683 "name": "BaseBdev3", 00:09:55.683 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.683 "is_configured": false, 00:09:55.683 "data_offset": 0, 00:09:55.683 "data_size": 0 00:09:55.683 }, 00:09:55.683 { 00:09:55.683 "name": "BaseBdev4", 00:09:55.683 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.683 "is_configured": false, 00:09:55.683 "data_offset": 0, 00:09:55.683 "data_size": 0 00:09:55.683 } 00:09:55.683 ] 00:09:55.683 }' 00:09:55.683 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.683 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.252 [2024-12-07 05:37:29.409646] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:56.252 BaseBdev3 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.252 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.253 [ 00:09:56.253 { 00:09:56.253 "name": "BaseBdev3", 00:09:56.253 "aliases": [ 00:09:56.253 "37f221ae-e193-4dbc-bcba-8086f8c10900" 00:09:56.253 ], 00:09:56.253 "product_name": "Malloc disk", 00:09:56.253 "block_size": 512, 00:09:56.253 "num_blocks": 65536, 00:09:56.253 "uuid": "37f221ae-e193-4dbc-bcba-8086f8c10900", 00:09:56.253 "assigned_rate_limits": { 00:09:56.253 "rw_ios_per_sec": 0, 00:09:56.253 "rw_mbytes_per_sec": 0, 00:09:56.253 "r_mbytes_per_sec": 0, 00:09:56.253 "w_mbytes_per_sec": 0 00:09:56.253 }, 00:09:56.253 "claimed": true, 00:09:56.253 "claim_type": "exclusive_write", 00:09:56.253 "zoned": false, 00:09:56.253 "supported_io_types": { 00:09:56.253 "read": true, 00:09:56.253 "write": true, 00:09:56.253 "unmap": true, 00:09:56.253 "flush": true, 00:09:56.253 "reset": true, 00:09:56.253 "nvme_admin": false, 00:09:56.253 "nvme_io": false, 00:09:56.253 "nvme_io_md": false, 00:09:56.253 "write_zeroes": true, 00:09:56.253 "zcopy": true, 00:09:56.253 "get_zone_info": false, 00:09:56.253 "zone_management": false, 00:09:56.253 "zone_append": false, 00:09:56.253 "compare": false, 00:09:56.253 "compare_and_write": false, 00:09:56.253 "abort": true, 00:09:56.253 "seek_hole": false, 00:09:56.253 "seek_data": false, 00:09:56.253 "copy": true, 00:09:56.253 "nvme_iov_md": false 00:09:56.253 }, 00:09:56.253 "memory_domains": [ 00:09:56.253 { 00:09:56.253 "dma_device_id": "system", 00:09:56.253 "dma_device_type": 1 00:09:56.253 }, 00:09:56.253 { 00:09:56.253 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.253 "dma_device_type": 2 00:09:56.253 } 00:09:56.253 ], 00:09:56.253 "driver_specific": {} 00:09:56.253 } 00:09:56.253 ] 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:56.253 "name": "Existed_Raid", 00:09:56.253 "uuid": "b720ebac-fe87-416c-8b96-f2271759dd3e", 00:09:56.253 "strip_size_kb": 64, 00:09:56.253 "state": "configuring", 00:09:56.253 "raid_level": "concat", 00:09:56.253 "superblock": true, 00:09:56.253 "num_base_bdevs": 4, 00:09:56.253 "num_base_bdevs_discovered": 3, 00:09:56.253 "num_base_bdevs_operational": 4, 00:09:56.253 "base_bdevs_list": [ 00:09:56.253 { 00:09:56.253 "name": "BaseBdev1", 00:09:56.253 "uuid": "ba09c5cb-0355-46d7-82dd-d911c416ca39", 00:09:56.253 "is_configured": true, 00:09:56.253 "data_offset": 2048, 00:09:56.253 "data_size": 63488 00:09:56.253 }, 00:09:56.253 { 00:09:56.253 "name": "BaseBdev2", 00:09:56.253 "uuid": "baa351a9-c8ea-42ee-b7a2-d178b1c7389c", 00:09:56.253 "is_configured": true, 00:09:56.253 "data_offset": 2048, 00:09:56.253 "data_size": 63488 00:09:56.253 }, 00:09:56.253 { 00:09:56.253 "name": "BaseBdev3", 00:09:56.253 "uuid": "37f221ae-e193-4dbc-bcba-8086f8c10900", 00:09:56.253 "is_configured": true, 00:09:56.253 "data_offset": 2048, 00:09:56.253 "data_size": 63488 00:09:56.253 }, 00:09:56.253 { 00:09:56.253 "name": "BaseBdev4", 00:09:56.253 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:56.253 "is_configured": false, 00:09:56.253 "data_offset": 0, 00:09:56.253 "data_size": 0 00:09:56.253 } 00:09:56.253 ] 00:09:56.253 }' 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:56.253 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.512 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:56.512 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.512 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.512 [2024-12-07 05:37:29.855880] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:56.512 [2024-12-07 05:37:29.856173] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:56.512 [2024-12-07 05:37:29.856225] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:56.512 BaseBdev4 00:09:56.512 [2024-12-07 05:37:29.856559] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:56.512 [2024-12-07 05:37:29.856743] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:56.512 [2024-12-07 05:37:29.856788] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:56.512 [2024-12-07 05:37:29.856951] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.513 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.513 [ 00:09:56.513 { 00:09:56.513 "name": "BaseBdev4", 00:09:56.513 "aliases": [ 00:09:56.513 "fe4e6510-8b79-4382-b95c-286f7c6e26f1" 00:09:56.513 ], 00:09:56.513 "product_name": "Malloc disk", 00:09:56.773 "block_size": 512, 00:09:56.773 "num_blocks": 65536, 00:09:56.773 "uuid": "fe4e6510-8b79-4382-b95c-286f7c6e26f1", 00:09:56.773 "assigned_rate_limits": { 00:09:56.773 "rw_ios_per_sec": 0, 00:09:56.773 "rw_mbytes_per_sec": 0, 00:09:56.773 "r_mbytes_per_sec": 0, 00:09:56.773 "w_mbytes_per_sec": 0 00:09:56.773 }, 00:09:56.773 "claimed": true, 00:09:56.773 "claim_type": "exclusive_write", 00:09:56.773 "zoned": false, 00:09:56.773 "supported_io_types": { 00:09:56.773 "read": true, 00:09:56.773 "write": true, 00:09:56.773 "unmap": true, 00:09:56.773 "flush": true, 00:09:56.773 "reset": true, 00:09:56.773 "nvme_admin": false, 00:09:56.773 "nvme_io": false, 00:09:56.773 "nvme_io_md": false, 00:09:56.773 "write_zeroes": true, 00:09:56.773 "zcopy": true, 00:09:56.773 "get_zone_info": false, 00:09:56.773 "zone_management": false, 00:09:56.773 "zone_append": false, 00:09:56.773 "compare": false, 00:09:56.773 "compare_and_write": false, 00:09:56.773 "abort": true, 00:09:56.773 "seek_hole": false, 00:09:56.773 "seek_data": false, 00:09:56.773 "copy": true, 00:09:56.773 "nvme_iov_md": false 00:09:56.773 }, 00:09:56.773 "memory_domains": [ 00:09:56.773 { 00:09:56.773 "dma_device_id": "system", 00:09:56.773 "dma_device_type": 1 00:09:56.773 }, 00:09:56.773 { 00:09:56.773 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.773 "dma_device_type": 2 00:09:56.773 } 00:09:56.773 ], 00:09:56.773 "driver_specific": {} 00:09:56.773 } 00:09:56.773 ] 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.773 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:56.773 "name": "Existed_Raid", 00:09:56.773 "uuid": "b720ebac-fe87-416c-8b96-f2271759dd3e", 00:09:56.773 "strip_size_kb": 64, 00:09:56.773 "state": "online", 00:09:56.773 "raid_level": "concat", 00:09:56.773 "superblock": true, 00:09:56.773 "num_base_bdevs": 4, 00:09:56.773 "num_base_bdevs_discovered": 4, 00:09:56.773 "num_base_bdevs_operational": 4, 00:09:56.773 "base_bdevs_list": [ 00:09:56.773 { 00:09:56.773 "name": "BaseBdev1", 00:09:56.773 "uuid": "ba09c5cb-0355-46d7-82dd-d911c416ca39", 00:09:56.774 "is_configured": true, 00:09:56.774 "data_offset": 2048, 00:09:56.774 "data_size": 63488 00:09:56.774 }, 00:09:56.774 { 00:09:56.774 "name": "BaseBdev2", 00:09:56.774 "uuid": "baa351a9-c8ea-42ee-b7a2-d178b1c7389c", 00:09:56.774 "is_configured": true, 00:09:56.774 "data_offset": 2048, 00:09:56.774 "data_size": 63488 00:09:56.774 }, 00:09:56.774 { 00:09:56.774 "name": "BaseBdev3", 00:09:56.774 "uuid": "37f221ae-e193-4dbc-bcba-8086f8c10900", 00:09:56.774 "is_configured": true, 00:09:56.774 "data_offset": 2048, 00:09:56.774 "data_size": 63488 00:09:56.774 }, 00:09:56.774 { 00:09:56.774 "name": "BaseBdev4", 00:09:56.774 "uuid": "fe4e6510-8b79-4382-b95c-286f7c6e26f1", 00:09:56.774 "is_configured": true, 00:09:56.774 "data_offset": 2048, 00:09:56.774 "data_size": 63488 00:09:56.774 } 00:09:56.774 ] 00:09:56.774 }' 00:09:56.774 05:37:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:56.774 05:37:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.034 [2024-12-07 05:37:30.319511] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.034 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:57.034 "name": "Existed_Raid", 00:09:57.034 "aliases": [ 00:09:57.034 "b720ebac-fe87-416c-8b96-f2271759dd3e" 00:09:57.034 ], 00:09:57.034 "product_name": "Raid Volume", 00:09:57.034 "block_size": 512, 00:09:57.034 "num_blocks": 253952, 00:09:57.034 "uuid": "b720ebac-fe87-416c-8b96-f2271759dd3e", 00:09:57.034 "assigned_rate_limits": { 00:09:57.034 "rw_ios_per_sec": 0, 00:09:57.034 "rw_mbytes_per_sec": 0, 00:09:57.034 "r_mbytes_per_sec": 0, 00:09:57.034 "w_mbytes_per_sec": 0 00:09:57.034 }, 00:09:57.034 "claimed": false, 00:09:57.034 "zoned": false, 00:09:57.034 "supported_io_types": { 00:09:57.034 "read": true, 00:09:57.034 "write": true, 00:09:57.034 "unmap": true, 00:09:57.034 "flush": true, 00:09:57.034 "reset": true, 00:09:57.034 "nvme_admin": false, 00:09:57.034 "nvme_io": false, 00:09:57.034 "nvme_io_md": false, 00:09:57.034 "write_zeroes": true, 00:09:57.034 "zcopy": false, 00:09:57.034 "get_zone_info": false, 00:09:57.034 "zone_management": false, 00:09:57.034 "zone_append": false, 00:09:57.034 "compare": false, 00:09:57.034 "compare_and_write": false, 00:09:57.034 "abort": false, 00:09:57.034 "seek_hole": false, 00:09:57.034 "seek_data": false, 00:09:57.034 "copy": false, 00:09:57.034 "nvme_iov_md": false 00:09:57.034 }, 00:09:57.034 "memory_domains": [ 00:09:57.034 { 00:09:57.034 "dma_device_id": "system", 00:09:57.034 "dma_device_type": 1 00:09:57.035 }, 00:09:57.035 { 00:09:57.035 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.035 "dma_device_type": 2 00:09:57.035 }, 00:09:57.035 { 00:09:57.035 "dma_device_id": "system", 00:09:57.035 "dma_device_type": 1 00:09:57.035 }, 00:09:57.035 { 00:09:57.035 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.035 "dma_device_type": 2 00:09:57.035 }, 00:09:57.035 { 00:09:57.035 "dma_device_id": "system", 00:09:57.035 "dma_device_type": 1 00:09:57.035 }, 00:09:57.035 { 00:09:57.035 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.035 "dma_device_type": 2 00:09:57.035 }, 00:09:57.035 { 00:09:57.035 "dma_device_id": "system", 00:09:57.035 "dma_device_type": 1 00:09:57.035 }, 00:09:57.035 { 00:09:57.035 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.035 "dma_device_type": 2 00:09:57.035 } 00:09:57.035 ], 00:09:57.035 "driver_specific": { 00:09:57.035 "raid": { 00:09:57.035 "uuid": "b720ebac-fe87-416c-8b96-f2271759dd3e", 00:09:57.035 "strip_size_kb": 64, 00:09:57.035 "state": "online", 00:09:57.035 "raid_level": "concat", 00:09:57.035 "superblock": true, 00:09:57.035 "num_base_bdevs": 4, 00:09:57.035 "num_base_bdevs_discovered": 4, 00:09:57.035 "num_base_bdevs_operational": 4, 00:09:57.035 "base_bdevs_list": [ 00:09:57.035 { 00:09:57.035 "name": "BaseBdev1", 00:09:57.035 "uuid": "ba09c5cb-0355-46d7-82dd-d911c416ca39", 00:09:57.035 "is_configured": true, 00:09:57.035 "data_offset": 2048, 00:09:57.035 "data_size": 63488 00:09:57.035 }, 00:09:57.035 { 00:09:57.035 "name": "BaseBdev2", 00:09:57.035 "uuid": "baa351a9-c8ea-42ee-b7a2-d178b1c7389c", 00:09:57.035 "is_configured": true, 00:09:57.035 "data_offset": 2048, 00:09:57.035 "data_size": 63488 00:09:57.035 }, 00:09:57.035 { 00:09:57.035 "name": "BaseBdev3", 00:09:57.035 "uuid": "37f221ae-e193-4dbc-bcba-8086f8c10900", 00:09:57.035 "is_configured": true, 00:09:57.035 "data_offset": 2048, 00:09:57.035 "data_size": 63488 00:09:57.035 }, 00:09:57.035 { 00:09:57.035 "name": "BaseBdev4", 00:09:57.035 "uuid": "fe4e6510-8b79-4382-b95c-286f7c6e26f1", 00:09:57.035 "is_configured": true, 00:09:57.035 "data_offset": 2048, 00:09:57.035 "data_size": 63488 00:09:57.035 } 00:09:57.035 ] 00:09:57.035 } 00:09:57.035 } 00:09:57.035 }' 00:09:57.035 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:57.035 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:57.035 BaseBdev2 00:09:57.035 BaseBdev3 00:09:57.035 BaseBdev4' 00:09:57.035 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.295 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.295 [2024-12-07 05:37:30.650644] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:57.295 [2024-12-07 05:37:30.650719] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:57.295 [2024-12-07 05:37:30.650835] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.555 "name": "Existed_Raid", 00:09:57.555 "uuid": "b720ebac-fe87-416c-8b96-f2271759dd3e", 00:09:57.555 "strip_size_kb": 64, 00:09:57.555 "state": "offline", 00:09:57.555 "raid_level": "concat", 00:09:57.555 "superblock": true, 00:09:57.555 "num_base_bdevs": 4, 00:09:57.555 "num_base_bdevs_discovered": 3, 00:09:57.555 "num_base_bdevs_operational": 3, 00:09:57.555 "base_bdevs_list": [ 00:09:57.555 { 00:09:57.555 "name": null, 00:09:57.555 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.555 "is_configured": false, 00:09:57.555 "data_offset": 0, 00:09:57.555 "data_size": 63488 00:09:57.555 }, 00:09:57.555 { 00:09:57.555 "name": "BaseBdev2", 00:09:57.555 "uuid": "baa351a9-c8ea-42ee-b7a2-d178b1c7389c", 00:09:57.555 "is_configured": true, 00:09:57.555 "data_offset": 2048, 00:09:57.555 "data_size": 63488 00:09:57.555 }, 00:09:57.555 { 00:09:57.555 "name": "BaseBdev3", 00:09:57.555 "uuid": "37f221ae-e193-4dbc-bcba-8086f8c10900", 00:09:57.555 "is_configured": true, 00:09:57.555 "data_offset": 2048, 00:09:57.555 "data_size": 63488 00:09:57.555 }, 00:09:57.555 { 00:09:57.555 "name": "BaseBdev4", 00:09:57.555 "uuid": "fe4e6510-8b79-4382-b95c-286f7c6e26f1", 00:09:57.555 "is_configured": true, 00:09:57.555 "data_offset": 2048, 00:09:57.555 "data_size": 63488 00:09:57.555 } 00:09:57.555 ] 00:09:57.555 }' 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.555 05:37:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.815 [2024-12-07 05:37:31.057459] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.815 [2024-12-07 05:37:31.124591] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.815 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.076 [2024-12-07 05:37:31.191761] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:58.076 [2024-12-07 05:37:31.191856] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.076 BaseBdev2 00:09:58.076 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.077 [ 00:09:58.077 { 00:09:58.077 "name": "BaseBdev2", 00:09:58.077 "aliases": [ 00:09:58.077 "f42288a1-296e-4dfe-bd17-dd7af42bb305" 00:09:58.077 ], 00:09:58.077 "product_name": "Malloc disk", 00:09:58.077 "block_size": 512, 00:09:58.077 "num_blocks": 65536, 00:09:58.077 "uuid": "f42288a1-296e-4dfe-bd17-dd7af42bb305", 00:09:58.077 "assigned_rate_limits": { 00:09:58.077 "rw_ios_per_sec": 0, 00:09:58.077 "rw_mbytes_per_sec": 0, 00:09:58.077 "r_mbytes_per_sec": 0, 00:09:58.077 "w_mbytes_per_sec": 0 00:09:58.077 }, 00:09:58.077 "claimed": false, 00:09:58.077 "zoned": false, 00:09:58.077 "supported_io_types": { 00:09:58.077 "read": true, 00:09:58.077 "write": true, 00:09:58.077 "unmap": true, 00:09:58.077 "flush": true, 00:09:58.077 "reset": true, 00:09:58.077 "nvme_admin": false, 00:09:58.077 "nvme_io": false, 00:09:58.077 "nvme_io_md": false, 00:09:58.077 "write_zeroes": true, 00:09:58.077 "zcopy": true, 00:09:58.077 "get_zone_info": false, 00:09:58.077 "zone_management": false, 00:09:58.077 "zone_append": false, 00:09:58.077 "compare": false, 00:09:58.077 "compare_and_write": false, 00:09:58.077 "abort": true, 00:09:58.077 "seek_hole": false, 00:09:58.077 "seek_data": false, 00:09:58.077 "copy": true, 00:09:58.077 "nvme_iov_md": false 00:09:58.077 }, 00:09:58.077 "memory_domains": [ 00:09:58.077 { 00:09:58.077 "dma_device_id": "system", 00:09:58.077 "dma_device_type": 1 00:09:58.077 }, 00:09:58.077 { 00:09:58.077 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.077 "dma_device_type": 2 00:09:58.077 } 00:09:58.077 ], 00:09:58.077 "driver_specific": {} 00:09:58.077 } 00:09:58.077 ] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.077 BaseBdev3 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.077 [ 00:09:58.077 { 00:09:58.077 "name": "BaseBdev3", 00:09:58.077 "aliases": [ 00:09:58.077 "cc131bc3-f5a2-4de8-921c-ae6d5178a74d" 00:09:58.077 ], 00:09:58.077 "product_name": "Malloc disk", 00:09:58.077 "block_size": 512, 00:09:58.077 "num_blocks": 65536, 00:09:58.077 "uuid": "cc131bc3-f5a2-4de8-921c-ae6d5178a74d", 00:09:58.077 "assigned_rate_limits": { 00:09:58.077 "rw_ios_per_sec": 0, 00:09:58.077 "rw_mbytes_per_sec": 0, 00:09:58.077 "r_mbytes_per_sec": 0, 00:09:58.077 "w_mbytes_per_sec": 0 00:09:58.077 }, 00:09:58.077 "claimed": false, 00:09:58.077 "zoned": false, 00:09:58.077 "supported_io_types": { 00:09:58.077 "read": true, 00:09:58.077 "write": true, 00:09:58.077 "unmap": true, 00:09:58.077 "flush": true, 00:09:58.077 "reset": true, 00:09:58.077 "nvme_admin": false, 00:09:58.077 "nvme_io": false, 00:09:58.077 "nvme_io_md": false, 00:09:58.077 "write_zeroes": true, 00:09:58.077 "zcopy": true, 00:09:58.077 "get_zone_info": false, 00:09:58.077 "zone_management": false, 00:09:58.077 "zone_append": false, 00:09:58.077 "compare": false, 00:09:58.077 "compare_and_write": false, 00:09:58.077 "abort": true, 00:09:58.077 "seek_hole": false, 00:09:58.077 "seek_data": false, 00:09:58.077 "copy": true, 00:09:58.077 "nvme_iov_md": false 00:09:58.077 }, 00:09:58.077 "memory_domains": [ 00:09:58.077 { 00:09:58.077 "dma_device_id": "system", 00:09:58.077 "dma_device_type": 1 00:09:58.077 }, 00:09:58.077 { 00:09:58.077 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.077 "dma_device_type": 2 00:09:58.077 } 00:09:58.077 ], 00:09:58.077 "driver_specific": {} 00:09:58.077 } 00:09:58.077 ] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.077 BaseBdev4 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.077 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.077 [ 00:09:58.077 { 00:09:58.077 "name": "BaseBdev4", 00:09:58.077 "aliases": [ 00:09:58.077 "00ed2c3a-2976-48cb-8347-16105336a739" 00:09:58.077 ], 00:09:58.077 "product_name": "Malloc disk", 00:09:58.077 "block_size": 512, 00:09:58.077 "num_blocks": 65536, 00:09:58.077 "uuid": "00ed2c3a-2976-48cb-8347-16105336a739", 00:09:58.077 "assigned_rate_limits": { 00:09:58.077 "rw_ios_per_sec": 0, 00:09:58.077 "rw_mbytes_per_sec": 0, 00:09:58.077 "r_mbytes_per_sec": 0, 00:09:58.077 "w_mbytes_per_sec": 0 00:09:58.077 }, 00:09:58.077 "claimed": false, 00:09:58.077 "zoned": false, 00:09:58.078 "supported_io_types": { 00:09:58.078 "read": true, 00:09:58.078 "write": true, 00:09:58.078 "unmap": true, 00:09:58.078 "flush": true, 00:09:58.078 "reset": true, 00:09:58.078 "nvme_admin": false, 00:09:58.078 "nvme_io": false, 00:09:58.078 "nvme_io_md": false, 00:09:58.078 "write_zeroes": true, 00:09:58.078 "zcopy": true, 00:09:58.078 "get_zone_info": false, 00:09:58.078 "zone_management": false, 00:09:58.078 "zone_append": false, 00:09:58.078 "compare": false, 00:09:58.078 "compare_and_write": false, 00:09:58.078 "abort": true, 00:09:58.078 "seek_hole": false, 00:09:58.078 "seek_data": false, 00:09:58.078 "copy": true, 00:09:58.078 "nvme_iov_md": false 00:09:58.078 }, 00:09:58.078 "memory_domains": [ 00:09:58.078 { 00:09:58.078 "dma_device_id": "system", 00:09:58.078 "dma_device_type": 1 00:09:58.078 }, 00:09:58.078 { 00:09:58.078 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.078 "dma_device_type": 2 00:09:58.078 } 00:09:58.078 ], 00:09:58.078 "driver_specific": {} 00:09:58.078 } 00:09:58.078 ] 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.078 [2024-12-07 05:37:31.417013] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:58.078 [2024-12-07 05:37:31.417097] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:58.078 [2024-12-07 05:37:31.417176] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:58.078 [2024-12-07 05:37:31.419148] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:58.078 [2024-12-07 05:37:31.419195] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.078 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.338 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.338 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.338 "name": "Existed_Raid", 00:09:58.338 "uuid": "6b6a46a0-ec15-489d-abcf-d71dc8911ac2", 00:09:58.338 "strip_size_kb": 64, 00:09:58.338 "state": "configuring", 00:09:58.338 "raid_level": "concat", 00:09:58.338 "superblock": true, 00:09:58.338 "num_base_bdevs": 4, 00:09:58.338 "num_base_bdevs_discovered": 3, 00:09:58.338 "num_base_bdevs_operational": 4, 00:09:58.338 "base_bdevs_list": [ 00:09:58.338 { 00:09:58.338 "name": "BaseBdev1", 00:09:58.338 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.338 "is_configured": false, 00:09:58.338 "data_offset": 0, 00:09:58.338 "data_size": 0 00:09:58.338 }, 00:09:58.338 { 00:09:58.338 "name": "BaseBdev2", 00:09:58.338 "uuid": "f42288a1-296e-4dfe-bd17-dd7af42bb305", 00:09:58.338 "is_configured": true, 00:09:58.338 "data_offset": 2048, 00:09:58.338 "data_size": 63488 00:09:58.338 }, 00:09:58.338 { 00:09:58.338 "name": "BaseBdev3", 00:09:58.338 "uuid": "cc131bc3-f5a2-4de8-921c-ae6d5178a74d", 00:09:58.338 "is_configured": true, 00:09:58.338 "data_offset": 2048, 00:09:58.338 "data_size": 63488 00:09:58.338 }, 00:09:58.338 { 00:09:58.338 "name": "BaseBdev4", 00:09:58.338 "uuid": "00ed2c3a-2976-48cb-8347-16105336a739", 00:09:58.338 "is_configured": true, 00:09:58.338 "data_offset": 2048, 00:09:58.338 "data_size": 63488 00:09:58.338 } 00:09:58.338 ] 00:09:58.338 }' 00:09:58.338 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.338 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.599 [2024-12-07 05:37:31.852277] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.599 "name": "Existed_Raid", 00:09:58.599 "uuid": "6b6a46a0-ec15-489d-abcf-d71dc8911ac2", 00:09:58.599 "strip_size_kb": 64, 00:09:58.599 "state": "configuring", 00:09:58.599 "raid_level": "concat", 00:09:58.599 "superblock": true, 00:09:58.599 "num_base_bdevs": 4, 00:09:58.599 "num_base_bdevs_discovered": 2, 00:09:58.599 "num_base_bdevs_operational": 4, 00:09:58.599 "base_bdevs_list": [ 00:09:58.599 { 00:09:58.599 "name": "BaseBdev1", 00:09:58.599 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.599 "is_configured": false, 00:09:58.599 "data_offset": 0, 00:09:58.599 "data_size": 0 00:09:58.599 }, 00:09:58.599 { 00:09:58.599 "name": null, 00:09:58.599 "uuid": "f42288a1-296e-4dfe-bd17-dd7af42bb305", 00:09:58.599 "is_configured": false, 00:09:58.599 "data_offset": 0, 00:09:58.599 "data_size": 63488 00:09:58.599 }, 00:09:58.599 { 00:09:58.599 "name": "BaseBdev3", 00:09:58.599 "uuid": "cc131bc3-f5a2-4de8-921c-ae6d5178a74d", 00:09:58.599 "is_configured": true, 00:09:58.599 "data_offset": 2048, 00:09:58.599 "data_size": 63488 00:09:58.599 }, 00:09:58.599 { 00:09:58.599 "name": "BaseBdev4", 00:09:58.599 "uuid": "00ed2c3a-2976-48cb-8347-16105336a739", 00:09:58.599 "is_configured": true, 00:09:58.599 "data_offset": 2048, 00:09:58.599 "data_size": 63488 00:09:58.599 } 00:09:58.599 ] 00:09:58.599 }' 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.599 05:37:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.168 [2024-12-07 05:37:32.350387] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:59.168 BaseBdev1 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.168 [ 00:09:59.168 { 00:09:59.168 "name": "BaseBdev1", 00:09:59.168 "aliases": [ 00:09:59.168 "94b5c7a3-acb1-4ba2-838c-0f6558b1aabb" 00:09:59.168 ], 00:09:59.168 "product_name": "Malloc disk", 00:09:59.168 "block_size": 512, 00:09:59.168 "num_blocks": 65536, 00:09:59.168 "uuid": "94b5c7a3-acb1-4ba2-838c-0f6558b1aabb", 00:09:59.168 "assigned_rate_limits": { 00:09:59.168 "rw_ios_per_sec": 0, 00:09:59.168 "rw_mbytes_per_sec": 0, 00:09:59.168 "r_mbytes_per_sec": 0, 00:09:59.168 "w_mbytes_per_sec": 0 00:09:59.168 }, 00:09:59.168 "claimed": true, 00:09:59.168 "claim_type": "exclusive_write", 00:09:59.168 "zoned": false, 00:09:59.168 "supported_io_types": { 00:09:59.168 "read": true, 00:09:59.168 "write": true, 00:09:59.168 "unmap": true, 00:09:59.168 "flush": true, 00:09:59.168 "reset": true, 00:09:59.168 "nvme_admin": false, 00:09:59.168 "nvme_io": false, 00:09:59.168 "nvme_io_md": false, 00:09:59.168 "write_zeroes": true, 00:09:59.168 "zcopy": true, 00:09:59.168 "get_zone_info": false, 00:09:59.168 "zone_management": false, 00:09:59.168 "zone_append": false, 00:09:59.168 "compare": false, 00:09:59.168 "compare_and_write": false, 00:09:59.168 "abort": true, 00:09:59.168 "seek_hole": false, 00:09:59.168 "seek_data": false, 00:09:59.168 "copy": true, 00:09:59.168 "nvme_iov_md": false 00:09:59.168 }, 00:09:59.168 "memory_domains": [ 00:09:59.168 { 00:09:59.168 "dma_device_id": "system", 00:09:59.168 "dma_device_type": 1 00:09:59.168 }, 00:09:59.168 { 00:09:59.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.168 "dma_device_type": 2 00:09:59.168 } 00:09:59.168 ], 00:09:59.168 "driver_specific": {} 00:09:59.168 } 00:09:59.168 ] 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.168 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.168 "name": "Existed_Raid", 00:09:59.168 "uuid": "6b6a46a0-ec15-489d-abcf-d71dc8911ac2", 00:09:59.168 "strip_size_kb": 64, 00:09:59.168 "state": "configuring", 00:09:59.169 "raid_level": "concat", 00:09:59.169 "superblock": true, 00:09:59.169 "num_base_bdevs": 4, 00:09:59.169 "num_base_bdevs_discovered": 3, 00:09:59.169 "num_base_bdevs_operational": 4, 00:09:59.169 "base_bdevs_list": [ 00:09:59.169 { 00:09:59.169 "name": "BaseBdev1", 00:09:59.169 "uuid": "94b5c7a3-acb1-4ba2-838c-0f6558b1aabb", 00:09:59.169 "is_configured": true, 00:09:59.169 "data_offset": 2048, 00:09:59.169 "data_size": 63488 00:09:59.169 }, 00:09:59.169 { 00:09:59.169 "name": null, 00:09:59.169 "uuid": "f42288a1-296e-4dfe-bd17-dd7af42bb305", 00:09:59.169 "is_configured": false, 00:09:59.169 "data_offset": 0, 00:09:59.169 "data_size": 63488 00:09:59.169 }, 00:09:59.169 { 00:09:59.169 "name": "BaseBdev3", 00:09:59.169 "uuid": "cc131bc3-f5a2-4de8-921c-ae6d5178a74d", 00:09:59.169 "is_configured": true, 00:09:59.169 "data_offset": 2048, 00:09:59.169 "data_size": 63488 00:09:59.169 }, 00:09:59.169 { 00:09:59.169 "name": "BaseBdev4", 00:09:59.169 "uuid": "00ed2c3a-2976-48cb-8347-16105336a739", 00:09:59.169 "is_configured": true, 00:09:59.169 "data_offset": 2048, 00:09:59.169 "data_size": 63488 00:09:59.169 } 00:09:59.169 ] 00:09:59.169 }' 00:09:59.169 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.169 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.738 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:59.738 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.738 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.739 [2024-12-07 05:37:32.889542] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.739 "name": "Existed_Raid", 00:09:59.739 "uuid": "6b6a46a0-ec15-489d-abcf-d71dc8911ac2", 00:09:59.739 "strip_size_kb": 64, 00:09:59.739 "state": "configuring", 00:09:59.739 "raid_level": "concat", 00:09:59.739 "superblock": true, 00:09:59.739 "num_base_bdevs": 4, 00:09:59.739 "num_base_bdevs_discovered": 2, 00:09:59.739 "num_base_bdevs_operational": 4, 00:09:59.739 "base_bdevs_list": [ 00:09:59.739 { 00:09:59.739 "name": "BaseBdev1", 00:09:59.739 "uuid": "94b5c7a3-acb1-4ba2-838c-0f6558b1aabb", 00:09:59.739 "is_configured": true, 00:09:59.739 "data_offset": 2048, 00:09:59.739 "data_size": 63488 00:09:59.739 }, 00:09:59.739 { 00:09:59.739 "name": null, 00:09:59.739 "uuid": "f42288a1-296e-4dfe-bd17-dd7af42bb305", 00:09:59.739 "is_configured": false, 00:09:59.739 "data_offset": 0, 00:09:59.739 "data_size": 63488 00:09:59.739 }, 00:09:59.739 { 00:09:59.739 "name": null, 00:09:59.739 "uuid": "cc131bc3-f5a2-4de8-921c-ae6d5178a74d", 00:09:59.739 "is_configured": false, 00:09:59.739 "data_offset": 0, 00:09:59.739 "data_size": 63488 00:09:59.739 }, 00:09:59.739 { 00:09:59.739 "name": "BaseBdev4", 00:09:59.739 "uuid": "00ed2c3a-2976-48cb-8347-16105336a739", 00:09:59.739 "is_configured": true, 00:09:59.739 "data_offset": 2048, 00:09:59.739 "data_size": 63488 00:09:59.739 } 00:09:59.739 ] 00:09:59.739 }' 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.739 05:37:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.999 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.999 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.999 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.999 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:59.999 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.261 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:00.261 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.262 [2024-12-07 05:37:33.380740] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.262 "name": "Existed_Raid", 00:10:00.262 "uuid": "6b6a46a0-ec15-489d-abcf-d71dc8911ac2", 00:10:00.262 "strip_size_kb": 64, 00:10:00.262 "state": "configuring", 00:10:00.262 "raid_level": "concat", 00:10:00.262 "superblock": true, 00:10:00.262 "num_base_bdevs": 4, 00:10:00.262 "num_base_bdevs_discovered": 3, 00:10:00.262 "num_base_bdevs_operational": 4, 00:10:00.262 "base_bdevs_list": [ 00:10:00.262 { 00:10:00.262 "name": "BaseBdev1", 00:10:00.262 "uuid": "94b5c7a3-acb1-4ba2-838c-0f6558b1aabb", 00:10:00.262 "is_configured": true, 00:10:00.262 "data_offset": 2048, 00:10:00.262 "data_size": 63488 00:10:00.262 }, 00:10:00.262 { 00:10:00.262 "name": null, 00:10:00.262 "uuid": "f42288a1-296e-4dfe-bd17-dd7af42bb305", 00:10:00.262 "is_configured": false, 00:10:00.262 "data_offset": 0, 00:10:00.262 "data_size": 63488 00:10:00.262 }, 00:10:00.262 { 00:10:00.262 "name": "BaseBdev3", 00:10:00.262 "uuid": "cc131bc3-f5a2-4de8-921c-ae6d5178a74d", 00:10:00.262 "is_configured": true, 00:10:00.262 "data_offset": 2048, 00:10:00.262 "data_size": 63488 00:10:00.262 }, 00:10:00.262 { 00:10:00.262 "name": "BaseBdev4", 00:10:00.262 "uuid": "00ed2c3a-2976-48cb-8347-16105336a739", 00:10:00.262 "is_configured": true, 00:10:00.262 "data_offset": 2048, 00:10:00.262 "data_size": 63488 00:10:00.262 } 00:10:00.262 ] 00:10:00.262 }' 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.262 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.521 [2024-12-07 05:37:33.828014] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.521 "name": "Existed_Raid", 00:10:00.521 "uuid": "6b6a46a0-ec15-489d-abcf-d71dc8911ac2", 00:10:00.521 "strip_size_kb": 64, 00:10:00.521 "state": "configuring", 00:10:00.521 "raid_level": "concat", 00:10:00.521 "superblock": true, 00:10:00.521 "num_base_bdevs": 4, 00:10:00.521 "num_base_bdevs_discovered": 2, 00:10:00.521 "num_base_bdevs_operational": 4, 00:10:00.521 "base_bdevs_list": [ 00:10:00.521 { 00:10:00.521 "name": null, 00:10:00.521 "uuid": "94b5c7a3-acb1-4ba2-838c-0f6558b1aabb", 00:10:00.521 "is_configured": false, 00:10:00.521 "data_offset": 0, 00:10:00.521 "data_size": 63488 00:10:00.521 }, 00:10:00.521 { 00:10:00.521 "name": null, 00:10:00.521 "uuid": "f42288a1-296e-4dfe-bd17-dd7af42bb305", 00:10:00.521 "is_configured": false, 00:10:00.521 "data_offset": 0, 00:10:00.521 "data_size": 63488 00:10:00.521 }, 00:10:00.521 { 00:10:00.521 "name": "BaseBdev3", 00:10:00.521 "uuid": "cc131bc3-f5a2-4de8-921c-ae6d5178a74d", 00:10:00.521 "is_configured": true, 00:10:00.521 "data_offset": 2048, 00:10:00.521 "data_size": 63488 00:10:00.521 }, 00:10:00.521 { 00:10:00.521 "name": "BaseBdev4", 00:10:00.521 "uuid": "00ed2c3a-2976-48cb-8347-16105336a739", 00:10:00.521 "is_configured": true, 00:10:00.521 "data_offset": 2048, 00:10:00.521 "data_size": 63488 00:10:00.521 } 00:10:00.521 ] 00:10:00.521 }' 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.521 05:37:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.090 [2024-12-07 05:37:34.309745] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.090 "name": "Existed_Raid", 00:10:01.090 "uuid": "6b6a46a0-ec15-489d-abcf-d71dc8911ac2", 00:10:01.090 "strip_size_kb": 64, 00:10:01.090 "state": "configuring", 00:10:01.090 "raid_level": "concat", 00:10:01.090 "superblock": true, 00:10:01.090 "num_base_bdevs": 4, 00:10:01.090 "num_base_bdevs_discovered": 3, 00:10:01.090 "num_base_bdevs_operational": 4, 00:10:01.090 "base_bdevs_list": [ 00:10:01.090 { 00:10:01.090 "name": null, 00:10:01.090 "uuid": "94b5c7a3-acb1-4ba2-838c-0f6558b1aabb", 00:10:01.090 "is_configured": false, 00:10:01.090 "data_offset": 0, 00:10:01.090 "data_size": 63488 00:10:01.090 }, 00:10:01.090 { 00:10:01.090 "name": "BaseBdev2", 00:10:01.090 "uuid": "f42288a1-296e-4dfe-bd17-dd7af42bb305", 00:10:01.090 "is_configured": true, 00:10:01.090 "data_offset": 2048, 00:10:01.090 "data_size": 63488 00:10:01.090 }, 00:10:01.090 { 00:10:01.090 "name": "BaseBdev3", 00:10:01.090 "uuid": "cc131bc3-f5a2-4de8-921c-ae6d5178a74d", 00:10:01.090 "is_configured": true, 00:10:01.090 "data_offset": 2048, 00:10:01.090 "data_size": 63488 00:10:01.090 }, 00:10:01.090 { 00:10:01.090 "name": "BaseBdev4", 00:10:01.090 "uuid": "00ed2c3a-2976-48cb-8347-16105336a739", 00:10:01.090 "is_configured": true, 00:10:01.090 "data_offset": 2048, 00:10:01.090 "data_size": 63488 00:10:01.090 } 00:10:01.090 ] 00:10:01.090 }' 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.090 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.350 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 94b5c7a3-acb1-4ba2-838c-0f6558b1aabb 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.610 NewBaseBdev 00:10:01.610 [2024-12-07 05:37:34.823847] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:01.610 [2024-12-07 05:37:34.824031] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:01.610 [2024-12-07 05:37:34.824044] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:01.610 [2024-12-07 05:37:34.824300] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:01.610 [2024-12-07 05:37:34.824408] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:01.610 [2024-12-07 05:37:34.824418] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:01.610 [2024-12-07 05:37:34.824509] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.610 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.610 [ 00:10:01.610 { 00:10:01.610 "name": "NewBaseBdev", 00:10:01.610 "aliases": [ 00:10:01.610 "94b5c7a3-acb1-4ba2-838c-0f6558b1aabb" 00:10:01.610 ], 00:10:01.610 "product_name": "Malloc disk", 00:10:01.610 "block_size": 512, 00:10:01.610 "num_blocks": 65536, 00:10:01.610 "uuid": "94b5c7a3-acb1-4ba2-838c-0f6558b1aabb", 00:10:01.610 "assigned_rate_limits": { 00:10:01.610 "rw_ios_per_sec": 0, 00:10:01.610 "rw_mbytes_per_sec": 0, 00:10:01.610 "r_mbytes_per_sec": 0, 00:10:01.610 "w_mbytes_per_sec": 0 00:10:01.610 }, 00:10:01.610 "claimed": true, 00:10:01.610 "claim_type": "exclusive_write", 00:10:01.610 "zoned": false, 00:10:01.610 "supported_io_types": { 00:10:01.610 "read": true, 00:10:01.610 "write": true, 00:10:01.610 "unmap": true, 00:10:01.610 "flush": true, 00:10:01.610 "reset": true, 00:10:01.610 "nvme_admin": false, 00:10:01.610 "nvme_io": false, 00:10:01.610 "nvme_io_md": false, 00:10:01.611 "write_zeroes": true, 00:10:01.611 "zcopy": true, 00:10:01.611 "get_zone_info": false, 00:10:01.611 "zone_management": false, 00:10:01.611 "zone_append": false, 00:10:01.611 "compare": false, 00:10:01.611 "compare_and_write": false, 00:10:01.611 "abort": true, 00:10:01.611 "seek_hole": false, 00:10:01.611 "seek_data": false, 00:10:01.611 "copy": true, 00:10:01.611 "nvme_iov_md": false 00:10:01.611 }, 00:10:01.611 "memory_domains": [ 00:10:01.611 { 00:10:01.611 "dma_device_id": "system", 00:10:01.611 "dma_device_type": 1 00:10:01.611 }, 00:10:01.611 { 00:10:01.611 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.611 "dma_device_type": 2 00:10:01.611 } 00:10:01.611 ], 00:10:01.611 "driver_specific": {} 00:10:01.611 } 00:10:01.611 ] 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.611 "name": "Existed_Raid", 00:10:01.611 "uuid": "6b6a46a0-ec15-489d-abcf-d71dc8911ac2", 00:10:01.611 "strip_size_kb": 64, 00:10:01.611 "state": "online", 00:10:01.611 "raid_level": "concat", 00:10:01.611 "superblock": true, 00:10:01.611 "num_base_bdevs": 4, 00:10:01.611 "num_base_bdevs_discovered": 4, 00:10:01.611 "num_base_bdevs_operational": 4, 00:10:01.611 "base_bdevs_list": [ 00:10:01.611 { 00:10:01.611 "name": "NewBaseBdev", 00:10:01.611 "uuid": "94b5c7a3-acb1-4ba2-838c-0f6558b1aabb", 00:10:01.611 "is_configured": true, 00:10:01.611 "data_offset": 2048, 00:10:01.611 "data_size": 63488 00:10:01.611 }, 00:10:01.611 { 00:10:01.611 "name": "BaseBdev2", 00:10:01.611 "uuid": "f42288a1-296e-4dfe-bd17-dd7af42bb305", 00:10:01.611 "is_configured": true, 00:10:01.611 "data_offset": 2048, 00:10:01.611 "data_size": 63488 00:10:01.611 }, 00:10:01.611 { 00:10:01.611 "name": "BaseBdev3", 00:10:01.611 "uuid": "cc131bc3-f5a2-4de8-921c-ae6d5178a74d", 00:10:01.611 "is_configured": true, 00:10:01.611 "data_offset": 2048, 00:10:01.611 "data_size": 63488 00:10:01.611 }, 00:10:01.611 { 00:10:01.611 "name": "BaseBdev4", 00:10:01.611 "uuid": "00ed2c3a-2976-48cb-8347-16105336a739", 00:10:01.611 "is_configured": true, 00:10:01.611 "data_offset": 2048, 00:10:01.611 "data_size": 63488 00:10:01.611 } 00:10:01.611 ] 00:10:01.611 }' 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.611 05:37:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:02.178 [2024-12-07 05:37:35.299434] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.178 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:02.178 "name": "Existed_Raid", 00:10:02.178 "aliases": [ 00:10:02.178 "6b6a46a0-ec15-489d-abcf-d71dc8911ac2" 00:10:02.178 ], 00:10:02.178 "product_name": "Raid Volume", 00:10:02.178 "block_size": 512, 00:10:02.178 "num_blocks": 253952, 00:10:02.178 "uuid": "6b6a46a0-ec15-489d-abcf-d71dc8911ac2", 00:10:02.178 "assigned_rate_limits": { 00:10:02.178 "rw_ios_per_sec": 0, 00:10:02.178 "rw_mbytes_per_sec": 0, 00:10:02.178 "r_mbytes_per_sec": 0, 00:10:02.178 "w_mbytes_per_sec": 0 00:10:02.178 }, 00:10:02.178 "claimed": false, 00:10:02.178 "zoned": false, 00:10:02.178 "supported_io_types": { 00:10:02.178 "read": true, 00:10:02.178 "write": true, 00:10:02.178 "unmap": true, 00:10:02.178 "flush": true, 00:10:02.178 "reset": true, 00:10:02.178 "nvme_admin": false, 00:10:02.178 "nvme_io": false, 00:10:02.178 "nvme_io_md": false, 00:10:02.178 "write_zeroes": true, 00:10:02.178 "zcopy": false, 00:10:02.178 "get_zone_info": false, 00:10:02.178 "zone_management": false, 00:10:02.178 "zone_append": false, 00:10:02.178 "compare": false, 00:10:02.178 "compare_and_write": false, 00:10:02.178 "abort": false, 00:10:02.178 "seek_hole": false, 00:10:02.178 "seek_data": false, 00:10:02.178 "copy": false, 00:10:02.178 "nvme_iov_md": false 00:10:02.178 }, 00:10:02.178 "memory_domains": [ 00:10:02.178 { 00:10:02.178 "dma_device_id": "system", 00:10:02.178 "dma_device_type": 1 00:10:02.178 }, 00:10:02.178 { 00:10:02.178 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.178 "dma_device_type": 2 00:10:02.178 }, 00:10:02.178 { 00:10:02.178 "dma_device_id": "system", 00:10:02.178 "dma_device_type": 1 00:10:02.178 }, 00:10:02.178 { 00:10:02.178 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.178 "dma_device_type": 2 00:10:02.178 }, 00:10:02.178 { 00:10:02.178 "dma_device_id": "system", 00:10:02.178 "dma_device_type": 1 00:10:02.178 }, 00:10:02.178 { 00:10:02.178 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.178 "dma_device_type": 2 00:10:02.178 }, 00:10:02.178 { 00:10:02.178 "dma_device_id": "system", 00:10:02.178 "dma_device_type": 1 00:10:02.178 }, 00:10:02.178 { 00:10:02.178 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.178 "dma_device_type": 2 00:10:02.178 } 00:10:02.178 ], 00:10:02.178 "driver_specific": { 00:10:02.178 "raid": { 00:10:02.178 "uuid": "6b6a46a0-ec15-489d-abcf-d71dc8911ac2", 00:10:02.178 "strip_size_kb": 64, 00:10:02.178 "state": "online", 00:10:02.178 "raid_level": "concat", 00:10:02.178 "superblock": true, 00:10:02.178 "num_base_bdevs": 4, 00:10:02.178 "num_base_bdevs_discovered": 4, 00:10:02.178 "num_base_bdevs_operational": 4, 00:10:02.179 "base_bdevs_list": [ 00:10:02.179 { 00:10:02.179 "name": "NewBaseBdev", 00:10:02.179 "uuid": "94b5c7a3-acb1-4ba2-838c-0f6558b1aabb", 00:10:02.179 "is_configured": true, 00:10:02.179 "data_offset": 2048, 00:10:02.179 "data_size": 63488 00:10:02.179 }, 00:10:02.179 { 00:10:02.179 "name": "BaseBdev2", 00:10:02.179 "uuid": "f42288a1-296e-4dfe-bd17-dd7af42bb305", 00:10:02.179 "is_configured": true, 00:10:02.179 "data_offset": 2048, 00:10:02.179 "data_size": 63488 00:10:02.179 }, 00:10:02.179 { 00:10:02.179 "name": "BaseBdev3", 00:10:02.179 "uuid": "cc131bc3-f5a2-4de8-921c-ae6d5178a74d", 00:10:02.179 "is_configured": true, 00:10:02.179 "data_offset": 2048, 00:10:02.179 "data_size": 63488 00:10:02.179 }, 00:10:02.179 { 00:10:02.179 "name": "BaseBdev4", 00:10:02.179 "uuid": "00ed2c3a-2976-48cb-8347-16105336a739", 00:10:02.179 "is_configured": true, 00:10:02.179 "data_offset": 2048, 00:10:02.179 "data_size": 63488 00:10:02.179 } 00:10:02.179 ] 00:10:02.179 } 00:10:02.179 } 00:10:02.179 }' 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:02.179 BaseBdev2 00:10:02.179 BaseBdev3 00:10:02.179 BaseBdev4' 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.179 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.437 [2024-12-07 05:37:35.598584] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:02.437 [2024-12-07 05:37:35.598670] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:02.437 [2024-12-07 05:37:35.598745] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:02.437 [2024-12-07 05:37:35.598815] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:02.437 [2024-12-07 05:37:35.598825] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 82470 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 82470 ']' 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 82470 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82470 00:10:02.437 killing process with pid 82470 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82470' 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 82470 00:10:02.437 [2024-12-07 05:37:35.645711] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:02.437 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 82470 00:10:02.437 [2024-12-07 05:37:35.686565] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:02.696 05:37:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:02.696 00:10:02.696 real 0m9.256s 00:10:02.696 user 0m15.948s 00:10:02.696 sys 0m1.785s 00:10:02.696 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:02.696 05:37:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.696 ************************************ 00:10:02.696 END TEST raid_state_function_test_sb 00:10:02.696 ************************************ 00:10:02.696 05:37:35 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:10:02.696 05:37:35 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:10:02.696 05:37:35 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:02.696 05:37:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:02.696 ************************************ 00:10:02.696 START TEST raid_superblock_test 00:10:02.696 ************************************ 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 4 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=83118 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 83118 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 83118 ']' 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:02.696 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:02.696 05:37:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.696 [2024-12-07 05:37:36.044408] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:10:02.696 [2024-12-07 05:37:36.044607] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83118 ] 00:10:02.968 [2024-12-07 05:37:36.201108] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:02.968 [2024-12-07 05:37:36.227968] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:02.968 [2024-12-07 05:37:36.270225] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:02.968 [2024-12-07 05:37:36.270341] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.534 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.794 malloc1 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.794 [2024-12-07 05:37:36.917756] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:03.794 [2024-12-07 05:37:36.917824] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:03.794 [2024-12-07 05:37:36.917847] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:03.794 [2024-12-07 05:37:36.917861] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:03.794 [2024-12-07 05:37:36.920040] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:03.794 [2024-12-07 05:37:36.920077] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:03.794 pt1 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.794 malloc2 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.794 [2024-12-07 05:37:36.946569] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:03.794 [2024-12-07 05:37:36.946689] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:03.794 [2024-12-07 05:37:36.946736] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:03.794 [2024-12-07 05:37:36.946770] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:03.794 [2024-12-07 05:37:36.949042] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:03.794 [2024-12-07 05:37:36.949109] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:03.794 pt2 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.794 malloc3 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.794 [2024-12-07 05:37:36.979260] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:03.794 [2024-12-07 05:37:36.979362] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:03.794 [2024-12-07 05:37:36.979408] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:03.794 [2024-12-07 05:37:36.979441] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:03.794 [2024-12-07 05:37:36.981708] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:03.794 [2024-12-07 05:37:36.981780] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:03.794 pt3 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.794 05:37:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.794 malloc4 00:10:03.794 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.794 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:03.794 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.794 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.794 [2024-12-07 05:37:37.019996] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:03.794 [2024-12-07 05:37:37.020049] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:03.794 [2024-12-07 05:37:37.020064] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:03.794 [2024-12-07 05:37:37.020076] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:03.794 [2024-12-07 05:37:37.022180] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:03.794 [2024-12-07 05:37:37.022218] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:03.794 pt4 00:10:03.794 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.794 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:03.794 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.795 [2024-12-07 05:37:37.032006] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:03.795 [2024-12-07 05:37:37.033803] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:03.795 [2024-12-07 05:37:37.033862] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:03.795 [2024-12-07 05:37:37.033926] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:03.795 [2024-12-07 05:37:37.034073] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:03.795 [2024-12-07 05:37:37.034086] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:03.795 [2024-12-07 05:37:37.034337] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:03.795 [2024-12-07 05:37:37.034480] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:03.795 [2024-12-07 05:37:37.034491] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:03.795 [2024-12-07 05:37:37.034602] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.795 "name": "raid_bdev1", 00:10:03.795 "uuid": "a4b11206-5d2e-4a61-bb5c-2b45f4ef146c", 00:10:03.795 "strip_size_kb": 64, 00:10:03.795 "state": "online", 00:10:03.795 "raid_level": "concat", 00:10:03.795 "superblock": true, 00:10:03.795 "num_base_bdevs": 4, 00:10:03.795 "num_base_bdevs_discovered": 4, 00:10:03.795 "num_base_bdevs_operational": 4, 00:10:03.795 "base_bdevs_list": [ 00:10:03.795 { 00:10:03.795 "name": "pt1", 00:10:03.795 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:03.795 "is_configured": true, 00:10:03.795 "data_offset": 2048, 00:10:03.795 "data_size": 63488 00:10:03.795 }, 00:10:03.795 { 00:10:03.795 "name": "pt2", 00:10:03.795 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:03.795 "is_configured": true, 00:10:03.795 "data_offset": 2048, 00:10:03.795 "data_size": 63488 00:10:03.795 }, 00:10:03.795 { 00:10:03.795 "name": "pt3", 00:10:03.795 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:03.795 "is_configured": true, 00:10:03.795 "data_offset": 2048, 00:10:03.795 "data_size": 63488 00:10:03.795 }, 00:10:03.795 { 00:10:03.795 "name": "pt4", 00:10:03.795 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:03.795 "is_configured": true, 00:10:03.795 "data_offset": 2048, 00:10:03.795 "data_size": 63488 00:10:03.795 } 00:10:03.795 ] 00:10:03.795 }' 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.795 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.365 [2024-12-07 05:37:37.515568] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.365 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:04.365 "name": "raid_bdev1", 00:10:04.365 "aliases": [ 00:10:04.365 "a4b11206-5d2e-4a61-bb5c-2b45f4ef146c" 00:10:04.365 ], 00:10:04.365 "product_name": "Raid Volume", 00:10:04.365 "block_size": 512, 00:10:04.365 "num_blocks": 253952, 00:10:04.365 "uuid": "a4b11206-5d2e-4a61-bb5c-2b45f4ef146c", 00:10:04.365 "assigned_rate_limits": { 00:10:04.365 "rw_ios_per_sec": 0, 00:10:04.366 "rw_mbytes_per_sec": 0, 00:10:04.366 "r_mbytes_per_sec": 0, 00:10:04.366 "w_mbytes_per_sec": 0 00:10:04.366 }, 00:10:04.366 "claimed": false, 00:10:04.366 "zoned": false, 00:10:04.366 "supported_io_types": { 00:10:04.366 "read": true, 00:10:04.366 "write": true, 00:10:04.366 "unmap": true, 00:10:04.366 "flush": true, 00:10:04.366 "reset": true, 00:10:04.366 "nvme_admin": false, 00:10:04.366 "nvme_io": false, 00:10:04.366 "nvme_io_md": false, 00:10:04.366 "write_zeroes": true, 00:10:04.366 "zcopy": false, 00:10:04.366 "get_zone_info": false, 00:10:04.366 "zone_management": false, 00:10:04.366 "zone_append": false, 00:10:04.366 "compare": false, 00:10:04.366 "compare_and_write": false, 00:10:04.366 "abort": false, 00:10:04.366 "seek_hole": false, 00:10:04.366 "seek_data": false, 00:10:04.366 "copy": false, 00:10:04.366 "nvme_iov_md": false 00:10:04.366 }, 00:10:04.366 "memory_domains": [ 00:10:04.366 { 00:10:04.366 "dma_device_id": "system", 00:10:04.366 "dma_device_type": 1 00:10:04.366 }, 00:10:04.366 { 00:10:04.366 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.366 "dma_device_type": 2 00:10:04.366 }, 00:10:04.366 { 00:10:04.366 "dma_device_id": "system", 00:10:04.366 "dma_device_type": 1 00:10:04.366 }, 00:10:04.366 { 00:10:04.366 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.366 "dma_device_type": 2 00:10:04.366 }, 00:10:04.366 { 00:10:04.366 "dma_device_id": "system", 00:10:04.366 "dma_device_type": 1 00:10:04.366 }, 00:10:04.366 { 00:10:04.366 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.366 "dma_device_type": 2 00:10:04.366 }, 00:10:04.366 { 00:10:04.366 "dma_device_id": "system", 00:10:04.366 "dma_device_type": 1 00:10:04.366 }, 00:10:04.366 { 00:10:04.366 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.366 "dma_device_type": 2 00:10:04.366 } 00:10:04.366 ], 00:10:04.366 "driver_specific": { 00:10:04.366 "raid": { 00:10:04.366 "uuid": "a4b11206-5d2e-4a61-bb5c-2b45f4ef146c", 00:10:04.366 "strip_size_kb": 64, 00:10:04.366 "state": "online", 00:10:04.366 "raid_level": "concat", 00:10:04.366 "superblock": true, 00:10:04.366 "num_base_bdevs": 4, 00:10:04.366 "num_base_bdevs_discovered": 4, 00:10:04.366 "num_base_bdevs_operational": 4, 00:10:04.366 "base_bdevs_list": [ 00:10:04.366 { 00:10:04.366 "name": "pt1", 00:10:04.366 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:04.366 "is_configured": true, 00:10:04.366 "data_offset": 2048, 00:10:04.366 "data_size": 63488 00:10:04.366 }, 00:10:04.366 { 00:10:04.366 "name": "pt2", 00:10:04.366 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:04.366 "is_configured": true, 00:10:04.366 "data_offset": 2048, 00:10:04.366 "data_size": 63488 00:10:04.366 }, 00:10:04.366 { 00:10:04.366 "name": "pt3", 00:10:04.366 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:04.366 "is_configured": true, 00:10:04.366 "data_offset": 2048, 00:10:04.366 "data_size": 63488 00:10:04.366 }, 00:10:04.366 { 00:10:04.366 "name": "pt4", 00:10:04.366 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:04.366 "is_configured": true, 00:10:04.366 "data_offset": 2048, 00:10:04.366 "data_size": 63488 00:10:04.366 } 00:10:04.366 ] 00:10:04.366 } 00:10:04.366 } 00:10:04.366 }' 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:04.366 pt2 00:10:04.366 pt3 00:10:04.366 pt4' 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.366 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.638 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.638 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.638 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.638 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.638 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:04.638 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.638 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.638 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.638 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.638 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.638 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.639 [2024-12-07 05:37:37.818920] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=a4b11206-5d2e-4a61-bb5c-2b45f4ef146c 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z a4b11206-5d2e-4a61-bb5c-2b45f4ef146c ']' 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.639 [2024-12-07 05:37:37.866534] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:04.639 [2024-12-07 05:37:37.866616] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:04.639 [2024-12-07 05:37:37.866728] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:04.639 [2024-12-07 05:37:37.866843] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:04.639 [2024-12-07 05:37:37.866895] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:04.639 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:04.912 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:04.912 05:37:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.912 [2024-12-07 05:37:38.010330] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:04.912 [2024-12-07 05:37:38.012401] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:04.912 [2024-12-07 05:37:38.012443] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:04.912 [2024-12-07 05:37:38.012473] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:04.912 [2024-12-07 05:37:38.012519] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:04.912 [2024-12-07 05:37:38.012565] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:04.912 [2024-12-07 05:37:38.012584] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:04.912 [2024-12-07 05:37:38.012601] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:04.912 [2024-12-07 05:37:38.012615] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:04.912 [2024-12-07 05:37:38.012709] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:10:04.912 request: 00:10:04.912 { 00:10:04.912 "name": "raid_bdev1", 00:10:04.912 "raid_level": "concat", 00:10:04.912 "base_bdevs": [ 00:10:04.912 "malloc1", 00:10:04.912 "malloc2", 00:10:04.912 "malloc3", 00:10:04.912 "malloc4" 00:10:04.912 ], 00:10:04.912 "strip_size_kb": 64, 00:10:04.912 "superblock": false, 00:10:04.912 "method": "bdev_raid_create", 00:10:04.912 "req_id": 1 00:10:04.912 } 00:10:04.912 Got JSON-RPC error response 00:10:04.912 response: 00:10:04.912 { 00:10:04.912 "code": -17, 00:10:04.912 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:04.912 } 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.912 [2024-12-07 05:37:38.078188] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:04.912 [2024-12-07 05:37:38.078242] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:04.912 [2024-12-07 05:37:38.078265] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:04.912 [2024-12-07 05:37:38.078274] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:04.912 [2024-12-07 05:37:38.080651] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:04.912 [2024-12-07 05:37:38.080682] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:04.912 [2024-12-07 05:37:38.080772] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:04.912 [2024-12-07 05:37:38.080827] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:04.912 pt1 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.912 "name": "raid_bdev1", 00:10:04.912 "uuid": "a4b11206-5d2e-4a61-bb5c-2b45f4ef146c", 00:10:04.912 "strip_size_kb": 64, 00:10:04.912 "state": "configuring", 00:10:04.912 "raid_level": "concat", 00:10:04.912 "superblock": true, 00:10:04.912 "num_base_bdevs": 4, 00:10:04.912 "num_base_bdevs_discovered": 1, 00:10:04.912 "num_base_bdevs_operational": 4, 00:10:04.912 "base_bdevs_list": [ 00:10:04.912 { 00:10:04.912 "name": "pt1", 00:10:04.912 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:04.912 "is_configured": true, 00:10:04.912 "data_offset": 2048, 00:10:04.912 "data_size": 63488 00:10:04.912 }, 00:10:04.912 { 00:10:04.912 "name": null, 00:10:04.912 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:04.912 "is_configured": false, 00:10:04.912 "data_offset": 2048, 00:10:04.912 "data_size": 63488 00:10:04.912 }, 00:10:04.912 { 00:10:04.912 "name": null, 00:10:04.912 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:04.912 "is_configured": false, 00:10:04.912 "data_offset": 2048, 00:10:04.912 "data_size": 63488 00:10:04.912 }, 00:10:04.912 { 00:10:04.912 "name": null, 00:10:04.912 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:04.912 "is_configured": false, 00:10:04.912 "data_offset": 2048, 00:10:04.912 "data_size": 63488 00:10:04.912 } 00:10:04.912 ] 00:10:04.912 }' 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.912 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.171 [2024-12-07 05:37:38.501483] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:05.171 [2024-12-07 05:37:38.501568] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:05.171 [2024-12-07 05:37:38.501590] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:10:05.171 [2024-12-07 05:37:38.501598] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:05.171 [2024-12-07 05:37:38.502034] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:05.171 [2024-12-07 05:37:38.502053] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:05.171 [2024-12-07 05:37:38.502133] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:05.171 [2024-12-07 05:37:38.502154] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:05.171 pt2 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.171 [2024-12-07 05:37:38.509480] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.171 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.430 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.430 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:05.430 "name": "raid_bdev1", 00:10:05.430 "uuid": "a4b11206-5d2e-4a61-bb5c-2b45f4ef146c", 00:10:05.430 "strip_size_kb": 64, 00:10:05.430 "state": "configuring", 00:10:05.430 "raid_level": "concat", 00:10:05.430 "superblock": true, 00:10:05.430 "num_base_bdevs": 4, 00:10:05.430 "num_base_bdevs_discovered": 1, 00:10:05.430 "num_base_bdevs_operational": 4, 00:10:05.430 "base_bdevs_list": [ 00:10:05.430 { 00:10:05.430 "name": "pt1", 00:10:05.430 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:05.430 "is_configured": true, 00:10:05.430 "data_offset": 2048, 00:10:05.430 "data_size": 63488 00:10:05.430 }, 00:10:05.430 { 00:10:05.430 "name": null, 00:10:05.430 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:05.430 "is_configured": false, 00:10:05.430 "data_offset": 0, 00:10:05.430 "data_size": 63488 00:10:05.430 }, 00:10:05.430 { 00:10:05.430 "name": null, 00:10:05.430 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:05.430 "is_configured": false, 00:10:05.430 "data_offset": 2048, 00:10:05.430 "data_size": 63488 00:10:05.430 }, 00:10:05.430 { 00:10:05.430 "name": null, 00:10:05.430 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:05.430 "is_configured": false, 00:10:05.430 "data_offset": 2048, 00:10:05.430 "data_size": 63488 00:10:05.430 } 00:10:05.430 ] 00:10:05.430 }' 00:10:05.430 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:05.430 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.690 [2024-12-07 05:37:38.924757] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:05.690 [2024-12-07 05:37:38.924880] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:05.690 [2024-12-07 05:37:38.924923] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:05.690 [2024-12-07 05:37:38.924955] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:05.690 [2024-12-07 05:37:38.925392] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:05.690 [2024-12-07 05:37:38.925449] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:05.690 [2024-12-07 05:37:38.925569] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:05.690 [2024-12-07 05:37:38.925655] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:05.690 pt2 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.690 [2024-12-07 05:37:38.936690] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:05.690 [2024-12-07 05:37:38.936773] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:05.690 [2024-12-07 05:37:38.936811] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:05.690 [2024-12-07 05:37:38.936842] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:05.690 [2024-12-07 05:37:38.937211] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:05.690 [2024-12-07 05:37:38.937268] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:05.690 [2024-12-07 05:37:38.937328] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:05.690 [2024-12-07 05:37:38.937351] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:05.690 pt3 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.690 [2024-12-07 05:37:38.948673] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:05.690 [2024-12-07 05:37:38.948760] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:05.690 [2024-12-07 05:37:38.948796] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:10:05.690 [2024-12-07 05:37:38.948825] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:05.690 [2024-12-07 05:37:38.949168] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:05.690 [2024-12-07 05:37:38.949240] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:05.690 [2024-12-07 05:37:38.949317] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:05.690 [2024-12-07 05:37:38.949363] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:05.690 [2024-12-07 05:37:38.949487] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:05.690 [2024-12-07 05:37:38.949525] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:05.690 [2024-12-07 05:37:38.949780] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:05.690 [2024-12-07 05:37:38.949937] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:05.690 [2024-12-07 05:37:38.949974] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:10:05.690 [2024-12-07 05:37:38.950108] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:05.690 pt4 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:05.690 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:05.691 "name": "raid_bdev1", 00:10:05.691 "uuid": "a4b11206-5d2e-4a61-bb5c-2b45f4ef146c", 00:10:05.691 "strip_size_kb": 64, 00:10:05.691 "state": "online", 00:10:05.691 "raid_level": "concat", 00:10:05.691 "superblock": true, 00:10:05.691 "num_base_bdevs": 4, 00:10:05.691 "num_base_bdevs_discovered": 4, 00:10:05.691 "num_base_bdevs_operational": 4, 00:10:05.691 "base_bdevs_list": [ 00:10:05.691 { 00:10:05.691 "name": "pt1", 00:10:05.691 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:05.691 "is_configured": true, 00:10:05.691 "data_offset": 2048, 00:10:05.691 "data_size": 63488 00:10:05.691 }, 00:10:05.691 { 00:10:05.691 "name": "pt2", 00:10:05.691 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:05.691 "is_configured": true, 00:10:05.691 "data_offset": 2048, 00:10:05.691 "data_size": 63488 00:10:05.691 }, 00:10:05.691 { 00:10:05.691 "name": "pt3", 00:10:05.691 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:05.691 "is_configured": true, 00:10:05.691 "data_offset": 2048, 00:10:05.691 "data_size": 63488 00:10:05.691 }, 00:10:05.691 { 00:10:05.691 "name": "pt4", 00:10:05.691 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:05.691 "is_configured": true, 00:10:05.691 "data_offset": 2048, 00:10:05.691 "data_size": 63488 00:10:05.691 } 00:10:05.691 ] 00:10:05.691 }' 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:05.691 05:37:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.260 [2024-12-07 05:37:39.352307] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.260 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:06.260 "name": "raid_bdev1", 00:10:06.260 "aliases": [ 00:10:06.260 "a4b11206-5d2e-4a61-bb5c-2b45f4ef146c" 00:10:06.260 ], 00:10:06.260 "product_name": "Raid Volume", 00:10:06.260 "block_size": 512, 00:10:06.260 "num_blocks": 253952, 00:10:06.260 "uuid": "a4b11206-5d2e-4a61-bb5c-2b45f4ef146c", 00:10:06.260 "assigned_rate_limits": { 00:10:06.260 "rw_ios_per_sec": 0, 00:10:06.260 "rw_mbytes_per_sec": 0, 00:10:06.260 "r_mbytes_per_sec": 0, 00:10:06.260 "w_mbytes_per_sec": 0 00:10:06.260 }, 00:10:06.260 "claimed": false, 00:10:06.260 "zoned": false, 00:10:06.260 "supported_io_types": { 00:10:06.260 "read": true, 00:10:06.260 "write": true, 00:10:06.260 "unmap": true, 00:10:06.260 "flush": true, 00:10:06.260 "reset": true, 00:10:06.260 "nvme_admin": false, 00:10:06.260 "nvme_io": false, 00:10:06.260 "nvme_io_md": false, 00:10:06.260 "write_zeroes": true, 00:10:06.260 "zcopy": false, 00:10:06.260 "get_zone_info": false, 00:10:06.260 "zone_management": false, 00:10:06.260 "zone_append": false, 00:10:06.260 "compare": false, 00:10:06.260 "compare_and_write": false, 00:10:06.260 "abort": false, 00:10:06.260 "seek_hole": false, 00:10:06.260 "seek_data": false, 00:10:06.260 "copy": false, 00:10:06.260 "nvme_iov_md": false 00:10:06.260 }, 00:10:06.260 "memory_domains": [ 00:10:06.260 { 00:10:06.260 "dma_device_id": "system", 00:10:06.260 "dma_device_type": 1 00:10:06.260 }, 00:10:06.260 { 00:10:06.260 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:06.260 "dma_device_type": 2 00:10:06.260 }, 00:10:06.260 { 00:10:06.260 "dma_device_id": "system", 00:10:06.260 "dma_device_type": 1 00:10:06.260 }, 00:10:06.261 { 00:10:06.261 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:06.261 "dma_device_type": 2 00:10:06.261 }, 00:10:06.261 { 00:10:06.261 "dma_device_id": "system", 00:10:06.261 "dma_device_type": 1 00:10:06.261 }, 00:10:06.261 { 00:10:06.261 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:06.261 "dma_device_type": 2 00:10:06.261 }, 00:10:06.261 { 00:10:06.261 "dma_device_id": "system", 00:10:06.261 "dma_device_type": 1 00:10:06.261 }, 00:10:06.261 { 00:10:06.261 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:06.261 "dma_device_type": 2 00:10:06.261 } 00:10:06.261 ], 00:10:06.261 "driver_specific": { 00:10:06.261 "raid": { 00:10:06.261 "uuid": "a4b11206-5d2e-4a61-bb5c-2b45f4ef146c", 00:10:06.261 "strip_size_kb": 64, 00:10:06.261 "state": "online", 00:10:06.261 "raid_level": "concat", 00:10:06.261 "superblock": true, 00:10:06.261 "num_base_bdevs": 4, 00:10:06.261 "num_base_bdevs_discovered": 4, 00:10:06.261 "num_base_bdevs_operational": 4, 00:10:06.261 "base_bdevs_list": [ 00:10:06.261 { 00:10:06.261 "name": "pt1", 00:10:06.261 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:06.261 "is_configured": true, 00:10:06.261 "data_offset": 2048, 00:10:06.261 "data_size": 63488 00:10:06.261 }, 00:10:06.261 { 00:10:06.261 "name": "pt2", 00:10:06.261 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:06.261 "is_configured": true, 00:10:06.261 "data_offset": 2048, 00:10:06.261 "data_size": 63488 00:10:06.261 }, 00:10:06.261 { 00:10:06.261 "name": "pt3", 00:10:06.261 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:06.261 "is_configured": true, 00:10:06.261 "data_offset": 2048, 00:10:06.261 "data_size": 63488 00:10:06.261 }, 00:10:06.261 { 00:10:06.261 "name": "pt4", 00:10:06.261 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:06.261 "is_configured": true, 00:10:06.261 "data_offset": 2048, 00:10:06.261 "data_size": 63488 00:10:06.261 } 00:10:06.261 ] 00:10:06.261 } 00:10:06.261 } 00:10:06.261 }' 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:06.261 pt2 00:10:06.261 pt3 00:10:06.261 pt4' 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.261 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.521 [2024-12-07 05:37:39.691738] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' a4b11206-5d2e-4a61-bb5c-2b45f4ef146c '!=' a4b11206-5d2e-4a61-bb5c-2b45f4ef146c ']' 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 83118 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 83118 ']' 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 83118 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83118 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:06.521 killing process with pid 83118 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83118' 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 83118 00:10:06.521 [2024-12-07 05:37:39.758043] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:06.521 [2024-12-07 05:37:39.758191] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:06.521 05:37:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 83118 00:10:06.521 [2024-12-07 05:37:39.758270] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:06.521 [2024-12-07 05:37:39.758282] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:10:06.521 [2024-12-07 05:37:39.802102] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:06.781 05:37:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:06.781 00:10:06.781 real 0m4.053s 00:10:06.781 user 0m6.463s 00:10:06.781 sys 0m0.829s 00:10:06.781 05:37:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:06.781 05:37:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.781 ************************************ 00:10:06.781 END TEST raid_superblock_test 00:10:06.781 ************************************ 00:10:06.781 05:37:40 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:10:06.781 05:37:40 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:06.781 05:37:40 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:06.781 05:37:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:06.781 ************************************ 00:10:06.781 START TEST raid_read_error_test 00:10:06.781 ************************************ 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 read 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Km1yGbDQ53 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83361 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83361 00:10:06.781 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 83361 ']' 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:06.781 05:37:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.041 [2024-12-07 05:37:40.183047] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:10:07.041 [2024-12-07 05:37:40.183173] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83361 ] 00:10:07.041 [2024-12-07 05:37:40.339993] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:07.041 [2024-12-07 05:37:40.364855] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:07.300 [2024-12-07 05:37:40.407129] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:07.300 [2024-12-07 05:37:40.407170] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 BaseBdev1_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 true 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 [2024-12-07 05:37:41.046146] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:07.870 [2024-12-07 05:37:41.046206] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:07.870 [2024-12-07 05:37:41.046240] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:07.870 [2024-12-07 05:37:41.046248] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:07.870 [2024-12-07 05:37:41.048443] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:07.870 [2024-12-07 05:37:41.048481] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:07.870 BaseBdev1 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 BaseBdev2_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 true 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 [2024-12-07 05:37:41.086671] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:07.870 [2024-12-07 05:37:41.086720] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:07.870 [2024-12-07 05:37:41.086736] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:07.870 [2024-12-07 05:37:41.086752] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:07.870 [2024-12-07 05:37:41.088820] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:07.870 [2024-12-07 05:37:41.088858] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:07.870 BaseBdev2 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 BaseBdev3_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 true 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 [2024-12-07 05:37:41.127080] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:07.870 [2024-12-07 05:37:41.127129] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:07.870 [2024-12-07 05:37:41.127147] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:07.870 [2024-12-07 05:37:41.127155] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:07.870 [2024-12-07 05:37:41.129213] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:07.870 [2024-12-07 05:37:41.129261] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:07.870 BaseBdev3 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 BaseBdev4_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 true 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 [2024-12-07 05:37:41.189639] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:07.870 [2024-12-07 05:37:41.189692] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:07.870 [2024-12-07 05:37:41.189716] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:07.870 [2024-12-07 05:37:41.189725] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:07.870 [2024-12-07 05:37:41.192049] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:07.870 [2024-12-07 05:37:41.192091] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:07.870 BaseBdev4 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.870 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.870 [2024-12-07 05:37:41.201643] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:07.870 [2024-12-07 05:37:41.203482] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:07.870 [2024-12-07 05:37:41.203563] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:07.870 [2024-12-07 05:37:41.203616] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:07.870 [2024-12-07 05:37:41.203834] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:07.870 [2024-12-07 05:37:41.203847] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:07.870 [2024-12-07 05:37:41.204110] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:10:07.870 [2024-12-07 05:37:41.204237] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:07.870 [2024-12-07 05:37:41.204249] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:07.871 [2024-12-07 05:37:41.204360] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.871 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.130 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:08.130 "name": "raid_bdev1", 00:10:08.130 "uuid": "fb06c069-c67a-4c15-9dce-38e6ca0303ec", 00:10:08.130 "strip_size_kb": 64, 00:10:08.130 "state": "online", 00:10:08.130 "raid_level": "concat", 00:10:08.130 "superblock": true, 00:10:08.130 "num_base_bdevs": 4, 00:10:08.130 "num_base_bdevs_discovered": 4, 00:10:08.130 "num_base_bdevs_operational": 4, 00:10:08.130 "base_bdevs_list": [ 00:10:08.130 { 00:10:08.130 "name": "BaseBdev1", 00:10:08.130 "uuid": "846a6ef1-a044-5848-a1a6-260f00e3e61d", 00:10:08.130 "is_configured": true, 00:10:08.130 "data_offset": 2048, 00:10:08.130 "data_size": 63488 00:10:08.130 }, 00:10:08.130 { 00:10:08.130 "name": "BaseBdev2", 00:10:08.130 "uuid": "ad01273f-865a-51fa-b250-1a37c96bf9f7", 00:10:08.130 "is_configured": true, 00:10:08.130 "data_offset": 2048, 00:10:08.130 "data_size": 63488 00:10:08.130 }, 00:10:08.130 { 00:10:08.130 "name": "BaseBdev3", 00:10:08.130 "uuid": "444ce59d-f70b-565b-a17d-e6af44ccd39e", 00:10:08.130 "is_configured": true, 00:10:08.130 "data_offset": 2048, 00:10:08.130 "data_size": 63488 00:10:08.130 }, 00:10:08.130 { 00:10:08.130 "name": "BaseBdev4", 00:10:08.130 "uuid": "9ff3bec8-e4c7-537c-ac6e-89343d34c3f1", 00:10:08.130 "is_configured": true, 00:10:08.130 "data_offset": 2048, 00:10:08.130 "data_size": 63488 00:10:08.130 } 00:10:08.130 ] 00:10:08.130 }' 00:10:08.130 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:08.130 05:37:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.388 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:08.388 05:37:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:08.388 [2024-12-07 05:37:41.741100] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.327 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.586 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.586 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:09.586 05:37:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.586 05:37:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.586 05:37:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.586 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.586 "name": "raid_bdev1", 00:10:09.586 "uuid": "fb06c069-c67a-4c15-9dce-38e6ca0303ec", 00:10:09.586 "strip_size_kb": 64, 00:10:09.586 "state": "online", 00:10:09.586 "raid_level": "concat", 00:10:09.586 "superblock": true, 00:10:09.586 "num_base_bdevs": 4, 00:10:09.586 "num_base_bdevs_discovered": 4, 00:10:09.586 "num_base_bdevs_operational": 4, 00:10:09.586 "base_bdevs_list": [ 00:10:09.586 { 00:10:09.586 "name": "BaseBdev1", 00:10:09.586 "uuid": "846a6ef1-a044-5848-a1a6-260f00e3e61d", 00:10:09.586 "is_configured": true, 00:10:09.586 "data_offset": 2048, 00:10:09.586 "data_size": 63488 00:10:09.586 }, 00:10:09.586 { 00:10:09.586 "name": "BaseBdev2", 00:10:09.586 "uuid": "ad01273f-865a-51fa-b250-1a37c96bf9f7", 00:10:09.586 "is_configured": true, 00:10:09.586 "data_offset": 2048, 00:10:09.586 "data_size": 63488 00:10:09.586 }, 00:10:09.586 { 00:10:09.586 "name": "BaseBdev3", 00:10:09.586 "uuid": "444ce59d-f70b-565b-a17d-e6af44ccd39e", 00:10:09.586 "is_configured": true, 00:10:09.586 "data_offset": 2048, 00:10:09.586 "data_size": 63488 00:10:09.586 }, 00:10:09.586 { 00:10:09.586 "name": "BaseBdev4", 00:10:09.586 "uuid": "9ff3bec8-e4c7-537c-ac6e-89343d34c3f1", 00:10:09.586 "is_configured": true, 00:10:09.586 "data_offset": 2048, 00:10:09.586 "data_size": 63488 00:10:09.586 } 00:10:09.586 ] 00:10:09.586 }' 00:10:09.586 05:37:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.586 05:37:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.845 05:37:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.846 [2024-12-07 05:37:43.113423] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:09.846 [2024-12-07 05:37:43.113515] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:09.846 [2024-12-07 05:37:43.116152] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:09.846 [2024-12-07 05:37:43.116247] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:09.846 [2024-12-07 05:37:43.116312] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:09.846 [2024-12-07 05:37:43.116371] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:09.846 { 00:10:09.846 "results": [ 00:10:09.846 { 00:10:09.846 "job": "raid_bdev1", 00:10:09.846 "core_mask": "0x1", 00:10:09.846 "workload": "randrw", 00:10:09.846 "percentage": 50, 00:10:09.846 "status": "finished", 00:10:09.846 "queue_depth": 1, 00:10:09.846 "io_size": 131072, 00:10:09.846 "runtime": 1.373184, 00:10:09.846 "iops": 15917.022045115586, 00:10:09.846 "mibps": 1989.6277556394482, 00:10:09.846 "io_failed": 1, 00:10:09.846 "io_timeout": 0, 00:10:09.846 "avg_latency_us": 86.86074923453924, 00:10:09.846 "min_latency_us": 26.606113537117903, 00:10:09.846 "max_latency_us": 1430.9170305676855 00:10:09.846 } 00:10:09.846 ], 00:10:09.846 "core_count": 1 00:10:09.846 } 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83361 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 83361 ']' 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 83361 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83361 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83361' 00:10:09.846 killing process with pid 83361 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 83361 00:10:09.846 [2024-12-07 05:37:43.165716] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:09.846 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 83361 00:10:09.846 [2024-12-07 05:37:43.200659] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:10.104 05:37:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:10.104 05:37:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Km1yGbDQ53 00:10:10.104 05:37:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:10.104 05:37:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:10:10.104 05:37:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:10.104 05:37:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:10.104 05:37:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:10.104 05:37:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:10:10.104 00:10:10.104 real 0m3.334s 00:10:10.104 user 0m4.218s 00:10:10.104 sys 0m0.503s 00:10:10.104 ************************************ 00:10:10.104 END TEST raid_read_error_test 00:10:10.104 ************************************ 00:10:10.104 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:10.104 05:37:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.364 05:37:43 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:10:10.364 05:37:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:10.364 05:37:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:10.364 05:37:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:10.364 ************************************ 00:10:10.364 START TEST raid_write_error_test 00:10:10.364 ************************************ 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 write 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:10.364 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.MlVisbzL7w 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83496 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83496 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 83496 ']' 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:10.365 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:10.365 05:37:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.365 [2024-12-07 05:37:43.585910] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:10:10.365 [2024-12-07 05:37:43.586022] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83496 ] 00:10:10.624 [2024-12-07 05:37:43.741967] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:10.624 [2024-12-07 05:37:43.767155] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:10.624 [2024-12-07 05:37:43.809153] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:10.624 [2024-12-07 05:37:43.809203] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.193 BaseBdev1_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.193 true 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.193 [2024-12-07 05:37:44.432182] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:11.193 [2024-12-07 05:37:44.432237] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.193 [2024-12-07 05:37:44.432263] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:11.193 [2024-12-07 05:37:44.432273] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.193 [2024-12-07 05:37:44.434492] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.193 [2024-12-07 05:37:44.434527] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:11.193 BaseBdev1 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.193 BaseBdev2_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.193 true 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.193 [2024-12-07 05:37:44.468655] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:11.193 [2024-12-07 05:37:44.468705] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.193 [2024-12-07 05:37:44.468723] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:11.193 [2024-12-07 05:37:44.468739] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.193 [2024-12-07 05:37:44.470887] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.193 [2024-12-07 05:37:44.470926] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:11.193 BaseBdev2 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.193 BaseBdev3_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.193 true 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.193 [2024-12-07 05:37:44.509255] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:11.193 [2024-12-07 05:37:44.509305] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.193 [2024-12-07 05:37:44.509341] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:11.193 [2024-12-07 05:37:44.509350] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.193 [2024-12-07 05:37:44.511501] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.193 [2024-12-07 05:37:44.511541] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:11.193 BaseBdev3 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.193 BaseBdev4_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.193 true 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.193 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.451 [2024-12-07 05:37:44.561676] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:11.451 [2024-12-07 05:37:44.561728] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.451 [2024-12-07 05:37:44.561766] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:11.451 [2024-12-07 05:37:44.561775] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.451 [2024-12-07 05:37:44.564030] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.451 [2024-12-07 05:37:44.564112] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:11.451 BaseBdev4 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.451 [2024-12-07 05:37:44.573698] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:11.451 [2024-12-07 05:37:44.575643] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:11.451 [2024-12-07 05:37:44.575727] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:11.451 [2024-12-07 05:37:44.575780] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:11.451 [2024-12-07 05:37:44.575981] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:11.451 [2024-12-07 05:37:44.576000] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:11.451 [2024-12-07 05:37:44.576268] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:10:11.451 [2024-12-07 05:37:44.576408] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:11.451 [2024-12-07 05:37:44.576425] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:11.451 [2024-12-07 05:37:44.576543] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.451 "name": "raid_bdev1", 00:10:11.451 "uuid": "6dd2e23c-8739-475b-8537-a03eddef5225", 00:10:11.451 "strip_size_kb": 64, 00:10:11.451 "state": "online", 00:10:11.451 "raid_level": "concat", 00:10:11.451 "superblock": true, 00:10:11.451 "num_base_bdevs": 4, 00:10:11.451 "num_base_bdevs_discovered": 4, 00:10:11.451 "num_base_bdevs_operational": 4, 00:10:11.451 "base_bdevs_list": [ 00:10:11.451 { 00:10:11.451 "name": "BaseBdev1", 00:10:11.451 "uuid": "5265b48e-832b-5071-b93b-7ba85cd24070", 00:10:11.451 "is_configured": true, 00:10:11.451 "data_offset": 2048, 00:10:11.451 "data_size": 63488 00:10:11.451 }, 00:10:11.451 { 00:10:11.451 "name": "BaseBdev2", 00:10:11.451 "uuid": "044d3c1c-6ce3-5b07-a061-511d9f356b6f", 00:10:11.451 "is_configured": true, 00:10:11.451 "data_offset": 2048, 00:10:11.451 "data_size": 63488 00:10:11.451 }, 00:10:11.451 { 00:10:11.451 "name": "BaseBdev3", 00:10:11.451 "uuid": "5a5eac24-5148-5251-9088-70802dee2791", 00:10:11.451 "is_configured": true, 00:10:11.451 "data_offset": 2048, 00:10:11.451 "data_size": 63488 00:10:11.451 }, 00:10:11.451 { 00:10:11.451 "name": "BaseBdev4", 00:10:11.451 "uuid": "dca4085f-8a2a-5c94-b2fe-f03148485be3", 00:10:11.451 "is_configured": true, 00:10:11.451 "data_offset": 2048, 00:10:11.451 "data_size": 63488 00:10:11.451 } 00:10:11.451 ] 00:10:11.451 }' 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.451 05:37:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.710 05:37:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:11.710 05:37:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:11.968 [2024-12-07 05:37:45.085162] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:12.902 "name": "raid_bdev1", 00:10:12.902 "uuid": "6dd2e23c-8739-475b-8537-a03eddef5225", 00:10:12.902 "strip_size_kb": 64, 00:10:12.902 "state": "online", 00:10:12.902 "raid_level": "concat", 00:10:12.902 "superblock": true, 00:10:12.902 "num_base_bdevs": 4, 00:10:12.902 "num_base_bdevs_discovered": 4, 00:10:12.902 "num_base_bdevs_operational": 4, 00:10:12.902 "base_bdevs_list": [ 00:10:12.902 { 00:10:12.902 "name": "BaseBdev1", 00:10:12.902 "uuid": "5265b48e-832b-5071-b93b-7ba85cd24070", 00:10:12.902 "is_configured": true, 00:10:12.902 "data_offset": 2048, 00:10:12.902 "data_size": 63488 00:10:12.902 }, 00:10:12.902 { 00:10:12.902 "name": "BaseBdev2", 00:10:12.902 "uuid": "044d3c1c-6ce3-5b07-a061-511d9f356b6f", 00:10:12.902 "is_configured": true, 00:10:12.902 "data_offset": 2048, 00:10:12.902 "data_size": 63488 00:10:12.902 }, 00:10:12.902 { 00:10:12.902 "name": "BaseBdev3", 00:10:12.902 "uuid": "5a5eac24-5148-5251-9088-70802dee2791", 00:10:12.902 "is_configured": true, 00:10:12.902 "data_offset": 2048, 00:10:12.902 "data_size": 63488 00:10:12.902 }, 00:10:12.902 { 00:10:12.902 "name": "BaseBdev4", 00:10:12.902 "uuid": "dca4085f-8a2a-5c94-b2fe-f03148485be3", 00:10:12.902 "is_configured": true, 00:10:12.902 "data_offset": 2048, 00:10:12.902 "data_size": 63488 00:10:12.902 } 00:10:12.902 ] 00:10:12.902 }' 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:12.902 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.160 [2024-12-07 05:37:46.485500] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:13.160 [2024-12-07 05:37:46.485563] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:13.160 [2024-12-07 05:37:46.490128] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:13.160 [2024-12-07 05:37:46.490334] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:13.160 [2024-12-07 05:37:46.490482] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:13.160 [2024-12-07 05:37:46.490567] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:13.160 { 00:10:13.160 "results": [ 00:10:13.160 { 00:10:13.160 "job": "raid_bdev1", 00:10:13.160 "core_mask": "0x1", 00:10:13.160 "workload": "randrw", 00:10:13.160 "percentage": 50, 00:10:13.160 "status": "finished", 00:10:13.160 "queue_depth": 1, 00:10:13.160 "io_size": 131072, 00:10:13.160 "runtime": 1.401431, 00:10:13.160 "iops": 15805.273324195055, 00:10:13.160 "mibps": 1975.659165524382, 00:10:13.160 "io_failed": 1, 00:10:13.160 "io_timeout": 0, 00:10:13.160 "avg_latency_us": 87.46911643958626, 00:10:13.160 "min_latency_us": 26.494323144104804, 00:10:13.160 "max_latency_us": 1445.2262008733624 00:10:13.160 } 00:10:13.160 ], 00:10:13.160 "core_count": 1 00:10:13.160 } 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83496 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 83496 ']' 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 83496 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83496 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83496' 00:10:13.160 killing process with pid 83496 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 83496 00:10:13.160 [2024-12-07 05:37:46.524799] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:13.160 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 83496 00:10:13.418 [2024-12-07 05:37:46.591635] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:13.676 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.MlVisbzL7w 00:10:13.676 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:13.676 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:13.676 ************************************ 00:10:13.676 END TEST raid_write_error_test 00:10:13.676 ************************************ 00:10:13.676 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.71 00:10:13.676 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:13.676 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:13.676 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:13.676 05:37:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.71 != \0\.\0\0 ]] 00:10:13.676 00:10:13.676 real 0m3.436s 00:10:13.676 user 0m4.315s 00:10:13.676 sys 0m0.496s 00:10:13.676 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:13.676 05:37:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.676 05:37:46 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:13.676 05:37:46 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:10:13.676 05:37:46 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:13.676 05:37:46 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:13.676 05:37:46 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:13.676 ************************************ 00:10:13.676 START TEST raid_state_function_test 00:10:13.676 ************************************ 00:10:13.676 05:37:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 false 00:10:13.676 05:37:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:13.676 05:37:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:13.676 05:37:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:13.676 05:37:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:13.676 05:37:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:13.676 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:13.677 Process raid pid: 83629 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=83629 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 83629' 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 83629 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 83629 ']' 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:13.677 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:13.677 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.936 [2024-12-07 05:37:47.087043] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:10:13.936 [2024-12-07 05:37:47.087252] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:13.936 [2024-12-07 05:37:47.240516] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:13.936 [2024-12-07 05:37:47.279543] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:14.195 [2024-12-07 05:37:47.355927] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:14.195 [2024-12-07 05:37:47.355975] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.762 [2024-12-07 05:37:47.934137] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:14.762 [2024-12-07 05:37:47.934308] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:14.762 [2024-12-07 05:37:47.934333] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:14.762 [2024-12-07 05:37:47.934347] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:14.762 [2024-12-07 05:37:47.934353] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:14.762 [2024-12-07 05:37:47.934366] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:14.762 [2024-12-07 05:37:47.934372] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:14.762 [2024-12-07 05:37:47.934382] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:14.762 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:14.763 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:14.763 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.763 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:14.763 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.763 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.763 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.763 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.763 "name": "Existed_Raid", 00:10:14.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.763 "strip_size_kb": 0, 00:10:14.763 "state": "configuring", 00:10:14.763 "raid_level": "raid1", 00:10:14.763 "superblock": false, 00:10:14.763 "num_base_bdevs": 4, 00:10:14.763 "num_base_bdevs_discovered": 0, 00:10:14.763 "num_base_bdevs_operational": 4, 00:10:14.763 "base_bdevs_list": [ 00:10:14.763 { 00:10:14.763 "name": "BaseBdev1", 00:10:14.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.763 "is_configured": false, 00:10:14.763 "data_offset": 0, 00:10:14.763 "data_size": 0 00:10:14.763 }, 00:10:14.763 { 00:10:14.763 "name": "BaseBdev2", 00:10:14.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.763 "is_configured": false, 00:10:14.763 "data_offset": 0, 00:10:14.763 "data_size": 0 00:10:14.763 }, 00:10:14.763 { 00:10:14.763 "name": "BaseBdev3", 00:10:14.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.763 "is_configured": false, 00:10:14.763 "data_offset": 0, 00:10:14.763 "data_size": 0 00:10:14.763 }, 00:10:14.763 { 00:10:14.763 "name": "BaseBdev4", 00:10:14.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.763 "is_configured": false, 00:10:14.763 "data_offset": 0, 00:10:14.763 "data_size": 0 00:10:14.763 } 00:10:14.763 ] 00:10:14.763 }' 00:10:14.763 05:37:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.763 05:37:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.033 [2024-12-07 05:37:48.341380] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:15.033 [2024-12-07 05:37:48.341514] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.033 [2024-12-07 05:37:48.353363] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:15.033 [2024-12-07 05:37:48.353464] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:15.033 [2024-12-07 05:37:48.353493] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:15.033 [2024-12-07 05:37:48.353515] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:15.033 [2024-12-07 05:37:48.353532] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:15.033 [2024-12-07 05:37:48.353552] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:15.033 [2024-12-07 05:37:48.353569] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:15.033 [2024-12-07 05:37:48.353590] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.033 [2024-12-07 05:37:48.380411] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:15.033 BaseBdev1 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.033 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.301 [ 00:10:15.301 { 00:10:15.301 "name": "BaseBdev1", 00:10:15.301 "aliases": [ 00:10:15.301 "44f1dc60-99e4-483e-b73d-5d4a56614336" 00:10:15.301 ], 00:10:15.301 "product_name": "Malloc disk", 00:10:15.301 "block_size": 512, 00:10:15.301 "num_blocks": 65536, 00:10:15.301 "uuid": "44f1dc60-99e4-483e-b73d-5d4a56614336", 00:10:15.301 "assigned_rate_limits": { 00:10:15.301 "rw_ios_per_sec": 0, 00:10:15.301 "rw_mbytes_per_sec": 0, 00:10:15.301 "r_mbytes_per_sec": 0, 00:10:15.301 "w_mbytes_per_sec": 0 00:10:15.301 }, 00:10:15.301 "claimed": true, 00:10:15.301 "claim_type": "exclusive_write", 00:10:15.301 "zoned": false, 00:10:15.301 "supported_io_types": { 00:10:15.301 "read": true, 00:10:15.301 "write": true, 00:10:15.301 "unmap": true, 00:10:15.301 "flush": true, 00:10:15.301 "reset": true, 00:10:15.301 "nvme_admin": false, 00:10:15.301 "nvme_io": false, 00:10:15.301 "nvme_io_md": false, 00:10:15.301 "write_zeroes": true, 00:10:15.301 "zcopy": true, 00:10:15.301 "get_zone_info": false, 00:10:15.301 "zone_management": false, 00:10:15.301 "zone_append": false, 00:10:15.301 "compare": false, 00:10:15.301 "compare_and_write": false, 00:10:15.301 "abort": true, 00:10:15.301 "seek_hole": false, 00:10:15.301 "seek_data": false, 00:10:15.301 "copy": true, 00:10:15.301 "nvme_iov_md": false 00:10:15.301 }, 00:10:15.301 "memory_domains": [ 00:10:15.301 { 00:10:15.301 "dma_device_id": "system", 00:10:15.301 "dma_device_type": 1 00:10:15.301 }, 00:10:15.301 { 00:10:15.301 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:15.301 "dma_device_type": 2 00:10:15.301 } 00:10:15.301 ], 00:10:15.301 "driver_specific": {} 00:10:15.301 } 00:10:15.301 ] 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.301 "name": "Existed_Raid", 00:10:15.301 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.301 "strip_size_kb": 0, 00:10:15.301 "state": "configuring", 00:10:15.301 "raid_level": "raid1", 00:10:15.301 "superblock": false, 00:10:15.301 "num_base_bdevs": 4, 00:10:15.301 "num_base_bdevs_discovered": 1, 00:10:15.301 "num_base_bdevs_operational": 4, 00:10:15.301 "base_bdevs_list": [ 00:10:15.301 { 00:10:15.301 "name": "BaseBdev1", 00:10:15.301 "uuid": "44f1dc60-99e4-483e-b73d-5d4a56614336", 00:10:15.301 "is_configured": true, 00:10:15.301 "data_offset": 0, 00:10:15.301 "data_size": 65536 00:10:15.301 }, 00:10:15.301 { 00:10:15.301 "name": "BaseBdev2", 00:10:15.301 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.301 "is_configured": false, 00:10:15.301 "data_offset": 0, 00:10:15.301 "data_size": 0 00:10:15.301 }, 00:10:15.301 { 00:10:15.301 "name": "BaseBdev3", 00:10:15.301 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.301 "is_configured": false, 00:10:15.301 "data_offset": 0, 00:10:15.301 "data_size": 0 00:10:15.301 }, 00:10:15.301 { 00:10:15.301 "name": "BaseBdev4", 00:10:15.301 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.301 "is_configured": false, 00:10:15.301 "data_offset": 0, 00:10:15.301 "data_size": 0 00:10:15.301 } 00:10:15.301 ] 00:10:15.301 }' 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.301 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.578 [2024-12-07 05:37:48.875685] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:15.578 [2024-12-07 05:37:48.875772] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.578 [2024-12-07 05:37:48.883676] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:15.578 [2024-12-07 05:37:48.885876] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:15.578 [2024-12-07 05:37:48.885922] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:15.578 [2024-12-07 05:37:48.885932] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:15.578 [2024-12-07 05:37:48.885941] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:15.578 [2024-12-07 05:37:48.885947] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:15.578 [2024-12-07 05:37:48.885955] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.578 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.578 "name": "Existed_Raid", 00:10:15.578 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.578 "strip_size_kb": 0, 00:10:15.578 "state": "configuring", 00:10:15.578 "raid_level": "raid1", 00:10:15.578 "superblock": false, 00:10:15.578 "num_base_bdevs": 4, 00:10:15.578 "num_base_bdevs_discovered": 1, 00:10:15.578 "num_base_bdevs_operational": 4, 00:10:15.578 "base_bdevs_list": [ 00:10:15.578 { 00:10:15.578 "name": "BaseBdev1", 00:10:15.578 "uuid": "44f1dc60-99e4-483e-b73d-5d4a56614336", 00:10:15.578 "is_configured": true, 00:10:15.578 "data_offset": 0, 00:10:15.578 "data_size": 65536 00:10:15.578 }, 00:10:15.578 { 00:10:15.578 "name": "BaseBdev2", 00:10:15.578 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.578 "is_configured": false, 00:10:15.578 "data_offset": 0, 00:10:15.578 "data_size": 0 00:10:15.578 }, 00:10:15.578 { 00:10:15.578 "name": "BaseBdev3", 00:10:15.579 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.579 "is_configured": false, 00:10:15.579 "data_offset": 0, 00:10:15.579 "data_size": 0 00:10:15.579 }, 00:10:15.579 { 00:10:15.579 "name": "BaseBdev4", 00:10:15.579 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.579 "is_configured": false, 00:10:15.579 "data_offset": 0, 00:10:15.579 "data_size": 0 00:10:15.579 } 00:10:15.579 ] 00:10:15.579 }' 00:10:15.579 05:37:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.579 05:37:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.148 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:16.148 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.148 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.148 [2024-12-07 05:37:49.363574] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:16.148 BaseBdev2 00:10:16.148 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.148 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:16.148 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.149 [ 00:10:16.149 { 00:10:16.149 "name": "BaseBdev2", 00:10:16.149 "aliases": [ 00:10:16.149 "aafdf5d6-c7d8-43a9-8dae-8a09e5a17811" 00:10:16.149 ], 00:10:16.149 "product_name": "Malloc disk", 00:10:16.149 "block_size": 512, 00:10:16.149 "num_blocks": 65536, 00:10:16.149 "uuid": "aafdf5d6-c7d8-43a9-8dae-8a09e5a17811", 00:10:16.149 "assigned_rate_limits": { 00:10:16.149 "rw_ios_per_sec": 0, 00:10:16.149 "rw_mbytes_per_sec": 0, 00:10:16.149 "r_mbytes_per_sec": 0, 00:10:16.149 "w_mbytes_per_sec": 0 00:10:16.149 }, 00:10:16.149 "claimed": true, 00:10:16.149 "claim_type": "exclusive_write", 00:10:16.149 "zoned": false, 00:10:16.149 "supported_io_types": { 00:10:16.149 "read": true, 00:10:16.149 "write": true, 00:10:16.149 "unmap": true, 00:10:16.149 "flush": true, 00:10:16.149 "reset": true, 00:10:16.149 "nvme_admin": false, 00:10:16.149 "nvme_io": false, 00:10:16.149 "nvme_io_md": false, 00:10:16.149 "write_zeroes": true, 00:10:16.149 "zcopy": true, 00:10:16.149 "get_zone_info": false, 00:10:16.149 "zone_management": false, 00:10:16.149 "zone_append": false, 00:10:16.149 "compare": false, 00:10:16.149 "compare_and_write": false, 00:10:16.149 "abort": true, 00:10:16.149 "seek_hole": false, 00:10:16.149 "seek_data": false, 00:10:16.149 "copy": true, 00:10:16.149 "nvme_iov_md": false 00:10:16.149 }, 00:10:16.149 "memory_domains": [ 00:10:16.149 { 00:10:16.149 "dma_device_id": "system", 00:10:16.149 "dma_device_type": 1 00:10:16.149 }, 00:10:16.149 { 00:10:16.149 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.149 "dma_device_type": 2 00:10:16.149 } 00:10:16.149 ], 00:10:16.149 "driver_specific": {} 00:10:16.149 } 00:10:16.149 ] 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:16.149 "name": "Existed_Raid", 00:10:16.149 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.149 "strip_size_kb": 0, 00:10:16.149 "state": "configuring", 00:10:16.149 "raid_level": "raid1", 00:10:16.149 "superblock": false, 00:10:16.149 "num_base_bdevs": 4, 00:10:16.149 "num_base_bdevs_discovered": 2, 00:10:16.149 "num_base_bdevs_operational": 4, 00:10:16.149 "base_bdevs_list": [ 00:10:16.149 { 00:10:16.149 "name": "BaseBdev1", 00:10:16.149 "uuid": "44f1dc60-99e4-483e-b73d-5d4a56614336", 00:10:16.149 "is_configured": true, 00:10:16.149 "data_offset": 0, 00:10:16.149 "data_size": 65536 00:10:16.149 }, 00:10:16.149 { 00:10:16.149 "name": "BaseBdev2", 00:10:16.149 "uuid": "aafdf5d6-c7d8-43a9-8dae-8a09e5a17811", 00:10:16.149 "is_configured": true, 00:10:16.149 "data_offset": 0, 00:10:16.149 "data_size": 65536 00:10:16.149 }, 00:10:16.149 { 00:10:16.149 "name": "BaseBdev3", 00:10:16.149 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.149 "is_configured": false, 00:10:16.149 "data_offset": 0, 00:10:16.149 "data_size": 0 00:10:16.149 }, 00:10:16.149 { 00:10:16.149 "name": "BaseBdev4", 00:10:16.149 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.149 "is_configured": false, 00:10:16.149 "data_offset": 0, 00:10:16.149 "data_size": 0 00:10:16.149 } 00:10:16.149 ] 00:10:16.149 }' 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:16.149 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.720 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.721 [2024-12-07 05:37:49.880772] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:16.721 BaseBdev3 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.721 [ 00:10:16.721 { 00:10:16.721 "name": "BaseBdev3", 00:10:16.721 "aliases": [ 00:10:16.721 "42408afa-dc59-4c36-95fa-99651768fe26" 00:10:16.721 ], 00:10:16.721 "product_name": "Malloc disk", 00:10:16.721 "block_size": 512, 00:10:16.721 "num_blocks": 65536, 00:10:16.721 "uuid": "42408afa-dc59-4c36-95fa-99651768fe26", 00:10:16.721 "assigned_rate_limits": { 00:10:16.721 "rw_ios_per_sec": 0, 00:10:16.721 "rw_mbytes_per_sec": 0, 00:10:16.721 "r_mbytes_per_sec": 0, 00:10:16.721 "w_mbytes_per_sec": 0 00:10:16.721 }, 00:10:16.721 "claimed": true, 00:10:16.721 "claim_type": "exclusive_write", 00:10:16.721 "zoned": false, 00:10:16.721 "supported_io_types": { 00:10:16.721 "read": true, 00:10:16.721 "write": true, 00:10:16.721 "unmap": true, 00:10:16.721 "flush": true, 00:10:16.721 "reset": true, 00:10:16.721 "nvme_admin": false, 00:10:16.721 "nvme_io": false, 00:10:16.721 "nvme_io_md": false, 00:10:16.721 "write_zeroes": true, 00:10:16.721 "zcopy": true, 00:10:16.721 "get_zone_info": false, 00:10:16.721 "zone_management": false, 00:10:16.721 "zone_append": false, 00:10:16.721 "compare": false, 00:10:16.721 "compare_and_write": false, 00:10:16.721 "abort": true, 00:10:16.721 "seek_hole": false, 00:10:16.721 "seek_data": false, 00:10:16.721 "copy": true, 00:10:16.721 "nvme_iov_md": false 00:10:16.721 }, 00:10:16.721 "memory_domains": [ 00:10:16.721 { 00:10:16.721 "dma_device_id": "system", 00:10:16.721 "dma_device_type": 1 00:10:16.721 }, 00:10:16.721 { 00:10:16.721 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.721 "dma_device_type": 2 00:10:16.721 } 00:10:16.721 ], 00:10:16.721 "driver_specific": {} 00:10:16.721 } 00:10:16.721 ] 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.721 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:16.721 "name": "Existed_Raid", 00:10:16.721 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.721 "strip_size_kb": 0, 00:10:16.721 "state": "configuring", 00:10:16.721 "raid_level": "raid1", 00:10:16.721 "superblock": false, 00:10:16.721 "num_base_bdevs": 4, 00:10:16.721 "num_base_bdevs_discovered": 3, 00:10:16.721 "num_base_bdevs_operational": 4, 00:10:16.721 "base_bdevs_list": [ 00:10:16.721 { 00:10:16.721 "name": "BaseBdev1", 00:10:16.721 "uuid": "44f1dc60-99e4-483e-b73d-5d4a56614336", 00:10:16.721 "is_configured": true, 00:10:16.721 "data_offset": 0, 00:10:16.721 "data_size": 65536 00:10:16.721 }, 00:10:16.721 { 00:10:16.721 "name": "BaseBdev2", 00:10:16.721 "uuid": "aafdf5d6-c7d8-43a9-8dae-8a09e5a17811", 00:10:16.721 "is_configured": true, 00:10:16.721 "data_offset": 0, 00:10:16.721 "data_size": 65536 00:10:16.721 }, 00:10:16.721 { 00:10:16.721 "name": "BaseBdev3", 00:10:16.721 "uuid": "42408afa-dc59-4c36-95fa-99651768fe26", 00:10:16.721 "is_configured": true, 00:10:16.722 "data_offset": 0, 00:10:16.722 "data_size": 65536 00:10:16.722 }, 00:10:16.722 { 00:10:16.722 "name": "BaseBdev4", 00:10:16.722 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.722 "is_configured": false, 00:10:16.722 "data_offset": 0, 00:10:16.722 "data_size": 0 00:10:16.722 } 00:10:16.722 ] 00:10:16.722 }' 00:10:16.722 05:37:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:16.722 05:37:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.289 [2024-12-07 05:37:50.382948] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:17.289 [2024-12-07 05:37:50.383003] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:17.289 [2024-12-07 05:37:50.383012] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:17.289 [2024-12-07 05:37:50.383303] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:17.289 [2024-12-07 05:37:50.383452] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:17.289 [2024-12-07 05:37:50.383464] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:17.289 [2024-12-07 05:37:50.383693] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:17.289 BaseBdev4 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.289 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.289 [ 00:10:17.289 { 00:10:17.289 "name": "BaseBdev4", 00:10:17.289 "aliases": [ 00:10:17.289 "3eacd251-586a-4879-a8a5-ff8f427bfad1" 00:10:17.289 ], 00:10:17.289 "product_name": "Malloc disk", 00:10:17.289 "block_size": 512, 00:10:17.289 "num_blocks": 65536, 00:10:17.289 "uuid": "3eacd251-586a-4879-a8a5-ff8f427bfad1", 00:10:17.289 "assigned_rate_limits": { 00:10:17.289 "rw_ios_per_sec": 0, 00:10:17.289 "rw_mbytes_per_sec": 0, 00:10:17.289 "r_mbytes_per_sec": 0, 00:10:17.289 "w_mbytes_per_sec": 0 00:10:17.289 }, 00:10:17.289 "claimed": true, 00:10:17.289 "claim_type": "exclusive_write", 00:10:17.289 "zoned": false, 00:10:17.289 "supported_io_types": { 00:10:17.289 "read": true, 00:10:17.289 "write": true, 00:10:17.289 "unmap": true, 00:10:17.289 "flush": true, 00:10:17.289 "reset": true, 00:10:17.289 "nvme_admin": false, 00:10:17.289 "nvme_io": false, 00:10:17.289 "nvme_io_md": false, 00:10:17.289 "write_zeroes": true, 00:10:17.289 "zcopy": true, 00:10:17.289 "get_zone_info": false, 00:10:17.289 "zone_management": false, 00:10:17.289 "zone_append": false, 00:10:17.289 "compare": false, 00:10:17.289 "compare_and_write": false, 00:10:17.289 "abort": true, 00:10:17.289 "seek_hole": false, 00:10:17.289 "seek_data": false, 00:10:17.289 "copy": true, 00:10:17.289 "nvme_iov_md": false 00:10:17.289 }, 00:10:17.289 "memory_domains": [ 00:10:17.289 { 00:10:17.289 "dma_device_id": "system", 00:10:17.289 "dma_device_type": 1 00:10:17.289 }, 00:10:17.289 { 00:10:17.289 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.289 "dma_device_type": 2 00:10:17.289 } 00:10:17.289 ], 00:10:17.289 "driver_specific": {} 00:10:17.289 } 00:10:17.290 ] 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:17.290 "name": "Existed_Raid", 00:10:17.290 "uuid": "f366aeb3-bd41-4e49-b570-5932bf8c2c16", 00:10:17.290 "strip_size_kb": 0, 00:10:17.290 "state": "online", 00:10:17.290 "raid_level": "raid1", 00:10:17.290 "superblock": false, 00:10:17.290 "num_base_bdevs": 4, 00:10:17.290 "num_base_bdevs_discovered": 4, 00:10:17.290 "num_base_bdevs_operational": 4, 00:10:17.290 "base_bdevs_list": [ 00:10:17.290 { 00:10:17.290 "name": "BaseBdev1", 00:10:17.290 "uuid": "44f1dc60-99e4-483e-b73d-5d4a56614336", 00:10:17.290 "is_configured": true, 00:10:17.290 "data_offset": 0, 00:10:17.290 "data_size": 65536 00:10:17.290 }, 00:10:17.290 { 00:10:17.290 "name": "BaseBdev2", 00:10:17.290 "uuid": "aafdf5d6-c7d8-43a9-8dae-8a09e5a17811", 00:10:17.290 "is_configured": true, 00:10:17.290 "data_offset": 0, 00:10:17.290 "data_size": 65536 00:10:17.290 }, 00:10:17.290 { 00:10:17.290 "name": "BaseBdev3", 00:10:17.290 "uuid": "42408afa-dc59-4c36-95fa-99651768fe26", 00:10:17.290 "is_configured": true, 00:10:17.290 "data_offset": 0, 00:10:17.290 "data_size": 65536 00:10:17.290 }, 00:10:17.290 { 00:10:17.290 "name": "BaseBdev4", 00:10:17.290 "uuid": "3eacd251-586a-4879-a8a5-ff8f427bfad1", 00:10:17.290 "is_configured": true, 00:10:17.290 "data_offset": 0, 00:10:17.290 "data_size": 65536 00:10:17.290 } 00:10:17.290 ] 00:10:17.290 }' 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:17.290 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:17.549 [2024-12-07 05:37:50.842609] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.549 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:17.549 "name": "Existed_Raid", 00:10:17.549 "aliases": [ 00:10:17.549 "f366aeb3-bd41-4e49-b570-5932bf8c2c16" 00:10:17.549 ], 00:10:17.549 "product_name": "Raid Volume", 00:10:17.549 "block_size": 512, 00:10:17.549 "num_blocks": 65536, 00:10:17.549 "uuid": "f366aeb3-bd41-4e49-b570-5932bf8c2c16", 00:10:17.549 "assigned_rate_limits": { 00:10:17.549 "rw_ios_per_sec": 0, 00:10:17.549 "rw_mbytes_per_sec": 0, 00:10:17.549 "r_mbytes_per_sec": 0, 00:10:17.549 "w_mbytes_per_sec": 0 00:10:17.549 }, 00:10:17.549 "claimed": false, 00:10:17.549 "zoned": false, 00:10:17.549 "supported_io_types": { 00:10:17.549 "read": true, 00:10:17.549 "write": true, 00:10:17.549 "unmap": false, 00:10:17.549 "flush": false, 00:10:17.549 "reset": true, 00:10:17.549 "nvme_admin": false, 00:10:17.549 "nvme_io": false, 00:10:17.549 "nvme_io_md": false, 00:10:17.549 "write_zeroes": true, 00:10:17.549 "zcopy": false, 00:10:17.549 "get_zone_info": false, 00:10:17.549 "zone_management": false, 00:10:17.550 "zone_append": false, 00:10:17.550 "compare": false, 00:10:17.550 "compare_and_write": false, 00:10:17.550 "abort": false, 00:10:17.550 "seek_hole": false, 00:10:17.550 "seek_data": false, 00:10:17.550 "copy": false, 00:10:17.550 "nvme_iov_md": false 00:10:17.550 }, 00:10:17.550 "memory_domains": [ 00:10:17.550 { 00:10:17.550 "dma_device_id": "system", 00:10:17.550 "dma_device_type": 1 00:10:17.550 }, 00:10:17.550 { 00:10:17.550 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.550 "dma_device_type": 2 00:10:17.550 }, 00:10:17.550 { 00:10:17.550 "dma_device_id": "system", 00:10:17.550 "dma_device_type": 1 00:10:17.550 }, 00:10:17.550 { 00:10:17.550 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.550 "dma_device_type": 2 00:10:17.550 }, 00:10:17.550 { 00:10:17.550 "dma_device_id": "system", 00:10:17.550 "dma_device_type": 1 00:10:17.550 }, 00:10:17.550 { 00:10:17.550 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.550 "dma_device_type": 2 00:10:17.550 }, 00:10:17.550 { 00:10:17.550 "dma_device_id": "system", 00:10:17.550 "dma_device_type": 1 00:10:17.550 }, 00:10:17.550 { 00:10:17.550 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.550 "dma_device_type": 2 00:10:17.550 } 00:10:17.550 ], 00:10:17.550 "driver_specific": { 00:10:17.550 "raid": { 00:10:17.550 "uuid": "f366aeb3-bd41-4e49-b570-5932bf8c2c16", 00:10:17.550 "strip_size_kb": 0, 00:10:17.550 "state": "online", 00:10:17.550 "raid_level": "raid1", 00:10:17.550 "superblock": false, 00:10:17.550 "num_base_bdevs": 4, 00:10:17.550 "num_base_bdevs_discovered": 4, 00:10:17.550 "num_base_bdevs_operational": 4, 00:10:17.550 "base_bdevs_list": [ 00:10:17.550 { 00:10:17.550 "name": "BaseBdev1", 00:10:17.550 "uuid": "44f1dc60-99e4-483e-b73d-5d4a56614336", 00:10:17.550 "is_configured": true, 00:10:17.550 "data_offset": 0, 00:10:17.550 "data_size": 65536 00:10:17.550 }, 00:10:17.550 { 00:10:17.550 "name": "BaseBdev2", 00:10:17.550 "uuid": "aafdf5d6-c7d8-43a9-8dae-8a09e5a17811", 00:10:17.550 "is_configured": true, 00:10:17.550 "data_offset": 0, 00:10:17.550 "data_size": 65536 00:10:17.550 }, 00:10:17.550 { 00:10:17.550 "name": "BaseBdev3", 00:10:17.550 "uuid": "42408afa-dc59-4c36-95fa-99651768fe26", 00:10:17.550 "is_configured": true, 00:10:17.550 "data_offset": 0, 00:10:17.550 "data_size": 65536 00:10:17.550 }, 00:10:17.550 { 00:10:17.550 "name": "BaseBdev4", 00:10:17.550 "uuid": "3eacd251-586a-4879-a8a5-ff8f427bfad1", 00:10:17.550 "is_configured": true, 00:10:17.550 "data_offset": 0, 00:10:17.550 "data_size": 65536 00:10:17.550 } 00:10:17.550 ] 00:10:17.550 } 00:10:17.550 } 00:10:17.550 }' 00:10:17.550 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:17.550 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:17.550 BaseBdev2 00:10:17.550 BaseBdev3 00:10:17.550 BaseBdev4' 00:10:17.809 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.810 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:17.810 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.810 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.810 05:37:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:17.810 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.810 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.810 05:37:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.810 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.810 [2024-12-07 05:37:51.173737] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.068 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:18.068 "name": "Existed_Raid", 00:10:18.068 "uuid": "f366aeb3-bd41-4e49-b570-5932bf8c2c16", 00:10:18.068 "strip_size_kb": 0, 00:10:18.068 "state": "online", 00:10:18.068 "raid_level": "raid1", 00:10:18.068 "superblock": false, 00:10:18.068 "num_base_bdevs": 4, 00:10:18.068 "num_base_bdevs_discovered": 3, 00:10:18.068 "num_base_bdevs_operational": 3, 00:10:18.068 "base_bdevs_list": [ 00:10:18.068 { 00:10:18.068 "name": null, 00:10:18.068 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.068 "is_configured": false, 00:10:18.068 "data_offset": 0, 00:10:18.068 "data_size": 65536 00:10:18.068 }, 00:10:18.069 { 00:10:18.069 "name": "BaseBdev2", 00:10:18.069 "uuid": "aafdf5d6-c7d8-43a9-8dae-8a09e5a17811", 00:10:18.069 "is_configured": true, 00:10:18.069 "data_offset": 0, 00:10:18.069 "data_size": 65536 00:10:18.069 }, 00:10:18.069 { 00:10:18.069 "name": "BaseBdev3", 00:10:18.069 "uuid": "42408afa-dc59-4c36-95fa-99651768fe26", 00:10:18.069 "is_configured": true, 00:10:18.069 "data_offset": 0, 00:10:18.069 "data_size": 65536 00:10:18.069 }, 00:10:18.069 { 00:10:18.069 "name": "BaseBdev4", 00:10:18.069 "uuid": "3eacd251-586a-4879-a8a5-ff8f427bfad1", 00:10:18.069 "is_configured": true, 00:10:18.069 "data_offset": 0, 00:10:18.069 "data_size": 65536 00:10:18.069 } 00:10:18.069 ] 00:10:18.069 }' 00:10:18.069 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:18.069 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.327 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:18.327 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:18.327 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.327 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:18.327 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.327 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.327 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.586 [2024-12-07 05:37:51.715886] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.586 [2024-12-07 05:37:51.774962] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.586 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.587 [2024-12-07 05:37:51.841933] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:18.587 [2024-12-07 05:37:51.842063] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:18.587 [2024-12-07 05:37:51.853473] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:18.587 [2024-12-07 05:37:51.853585] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:18.587 [2024-12-07 05:37:51.853603] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.587 BaseBdev2 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.587 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.587 [ 00:10:18.587 { 00:10:18.587 "name": "BaseBdev2", 00:10:18.587 "aliases": [ 00:10:18.587 "256ebd08-1af6-4530-a57a-b200efff626e" 00:10:18.587 ], 00:10:18.587 "product_name": "Malloc disk", 00:10:18.587 "block_size": 512, 00:10:18.587 "num_blocks": 65536, 00:10:18.587 "uuid": "256ebd08-1af6-4530-a57a-b200efff626e", 00:10:18.587 "assigned_rate_limits": { 00:10:18.587 "rw_ios_per_sec": 0, 00:10:18.587 "rw_mbytes_per_sec": 0, 00:10:18.587 "r_mbytes_per_sec": 0, 00:10:18.587 "w_mbytes_per_sec": 0 00:10:18.587 }, 00:10:18.587 "claimed": false, 00:10:18.587 "zoned": false, 00:10:18.587 "supported_io_types": { 00:10:18.587 "read": true, 00:10:18.587 "write": true, 00:10:18.587 "unmap": true, 00:10:18.587 "flush": true, 00:10:18.587 "reset": true, 00:10:18.587 "nvme_admin": false, 00:10:18.587 "nvme_io": false, 00:10:18.587 "nvme_io_md": false, 00:10:18.587 "write_zeroes": true, 00:10:18.587 "zcopy": true, 00:10:18.587 "get_zone_info": false, 00:10:18.587 "zone_management": false, 00:10:18.587 "zone_append": false, 00:10:18.587 "compare": false, 00:10:18.587 "compare_and_write": false, 00:10:18.587 "abort": true, 00:10:18.587 "seek_hole": false, 00:10:18.587 "seek_data": false, 00:10:18.587 "copy": true, 00:10:18.587 "nvme_iov_md": false 00:10:18.587 }, 00:10:18.587 "memory_domains": [ 00:10:18.587 { 00:10:18.587 "dma_device_id": "system", 00:10:18.587 "dma_device_type": 1 00:10:18.587 }, 00:10:18.846 { 00:10:18.846 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.846 "dma_device_type": 2 00:10:18.846 } 00:10:18.846 ], 00:10:18.846 "driver_specific": {} 00:10:18.846 } 00:10:18.846 ] 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.846 BaseBdev3 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:18.846 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:18.847 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:18.847 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.847 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.847 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.847 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:18.847 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.847 05:37:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.847 [ 00:10:18.847 { 00:10:18.847 "name": "BaseBdev3", 00:10:18.847 "aliases": [ 00:10:18.847 "3fa64264-5aba-4014-9c73-8c2581033a5c" 00:10:18.847 ], 00:10:18.847 "product_name": "Malloc disk", 00:10:18.847 "block_size": 512, 00:10:18.847 "num_blocks": 65536, 00:10:18.847 "uuid": "3fa64264-5aba-4014-9c73-8c2581033a5c", 00:10:18.847 "assigned_rate_limits": { 00:10:18.847 "rw_ios_per_sec": 0, 00:10:18.847 "rw_mbytes_per_sec": 0, 00:10:18.847 "r_mbytes_per_sec": 0, 00:10:18.847 "w_mbytes_per_sec": 0 00:10:18.847 }, 00:10:18.847 "claimed": false, 00:10:18.847 "zoned": false, 00:10:18.847 "supported_io_types": { 00:10:18.847 "read": true, 00:10:18.847 "write": true, 00:10:18.847 "unmap": true, 00:10:18.847 "flush": true, 00:10:18.847 "reset": true, 00:10:18.847 "nvme_admin": false, 00:10:18.847 "nvme_io": false, 00:10:18.847 "nvme_io_md": false, 00:10:18.847 "write_zeroes": true, 00:10:18.847 "zcopy": true, 00:10:18.847 "get_zone_info": false, 00:10:18.847 "zone_management": false, 00:10:18.847 "zone_append": false, 00:10:18.847 "compare": false, 00:10:18.847 "compare_and_write": false, 00:10:18.847 "abort": true, 00:10:18.847 "seek_hole": false, 00:10:18.847 "seek_data": false, 00:10:18.847 "copy": true, 00:10:18.847 "nvme_iov_md": false 00:10:18.847 }, 00:10:18.847 "memory_domains": [ 00:10:18.847 { 00:10:18.847 "dma_device_id": "system", 00:10:18.847 "dma_device_type": 1 00:10:18.847 }, 00:10:18.847 { 00:10:18.847 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.847 "dma_device_type": 2 00:10:18.847 } 00:10:18.847 ], 00:10:18.847 "driver_specific": {} 00:10:18.847 } 00:10:18.847 ] 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.847 BaseBdev4 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.847 [ 00:10:18.847 { 00:10:18.847 "name": "BaseBdev4", 00:10:18.847 "aliases": [ 00:10:18.847 "52635a23-96a4-4463-86a1-d4b1d12096ff" 00:10:18.847 ], 00:10:18.847 "product_name": "Malloc disk", 00:10:18.847 "block_size": 512, 00:10:18.847 "num_blocks": 65536, 00:10:18.847 "uuid": "52635a23-96a4-4463-86a1-d4b1d12096ff", 00:10:18.847 "assigned_rate_limits": { 00:10:18.847 "rw_ios_per_sec": 0, 00:10:18.847 "rw_mbytes_per_sec": 0, 00:10:18.847 "r_mbytes_per_sec": 0, 00:10:18.847 "w_mbytes_per_sec": 0 00:10:18.847 }, 00:10:18.847 "claimed": false, 00:10:18.847 "zoned": false, 00:10:18.847 "supported_io_types": { 00:10:18.847 "read": true, 00:10:18.847 "write": true, 00:10:18.847 "unmap": true, 00:10:18.847 "flush": true, 00:10:18.847 "reset": true, 00:10:18.847 "nvme_admin": false, 00:10:18.847 "nvme_io": false, 00:10:18.847 "nvme_io_md": false, 00:10:18.847 "write_zeroes": true, 00:10:18.847 "zcopy": true, 00:10:18.847 "get_zone_info": false, 00:10:18.847 "zone_management": false, 00:10:18.847 "zone_append": false, 00:10:18.847 "compare": false, 00:10:18.847 "compare_and_write": false, 00:10:18.847 "abort": true, 00:10:18.847 "seek_hole": false, 00:10:18.847 "seek_data": false, 00:10:18.847 "copy": true, 00:10:18.847 "nvme_iov_md": false 00:10:18.847 }, 00:10:18.847 "memory_domains": [ 00:10:18.847 { 00:10:18.847 "dma_device_id": "system", 00:10:18.847 "dma_device_type": 1 00:10:18.847 }, 00:10:18.847 { 00:10:18.847 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.847 "dma_device_type": 2 00:10:18.847 } 00:10:18.847 ], 00:10:18.847 "driver_specific": {} 00:10:18.847 } 00:10:18.847 ] 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.847 [2024-12-07 05:37:52.066210] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:18.847 [2024-12-07 05:37:52.066296] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:18.847 [2024-12-07 05:37:52.066336] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:18.847 [2024-12-07 05:37:52.068224] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:18.847 [2024-12-07 05:37:52.068318] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.847 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:18.847 "name": "Existed_Raid", 00:10:18.847 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.847 "strip_size_kb": 0, 00:10:18.847 "state": "configuring", 00:10:18.847 "raid_level": "raid1", 00:10:18.847 "superblock": false, 00:10:18.847 "num_base_bdevs": 4, 00:10:18.847 "num_base_bdevs_discovered": 3, 00:10:18.847 "num_base_bdevs_operational": 4, 00:10:18.847 "base_bdevs_list": [ 00:10:18.847 { 00:10:18.847 "name": "BaseBdev1", 00:10:18.847 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.847 "is_configured": false, 00:10:18.847 "data_offset": 0, 00:10:18.847 "data_size": 0 00:10:18.847 }, 00:10:18.847 { 00:10:18.847 "name": "BaseBdev2", 00:10:18.847 "uuid": "256ebd08-1af6-4530-a57a-b200efff626e", 00:10:18.847 "is_configured": true, 00:10:18.847 "data_offset": 0, 00:10:18.847 "data_size": 65536 00:10:18.847 }, 00:10:18.847 { 00:10:18.847 "name": "BaseBdev3", 00:10:18.847 "uuid": "3fa64264-5aba-4014-9c73-8c2581033a5c", 00:10:18.847 "is_configured": true, 00:10:18.847 "data_offset": 0, 00:10:18.848 "data_size": 65536 00:10:18.848 }, 00:10:18.848 { 00:10:18.848 "name": "BaseBdev4", 00:10:18.848 "uuid": "52635a23-96a4-4463-86a1-d4b1d12096ff", 00:10:18.848 "is_configured": true, 00:10:18.848 "data_offset": 0, 00:10:18.848 "data_size": 65536 00:10:18.848 } 00:10:18.848 ] 00:10:18.848 }' 00:10:18.848 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:18.848 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.107 [2024-12-07 05:37:52.457546] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.107 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.365 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.365 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.365 "name": "Existed_Raid", 00:10:19.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.365 "strip_size_kb": 0, 00:10:19.365 "state": "configuring", 00:10:19.365 "raid_level": "raid1", 00:10:19.365 "superblock": false, 00:10:19.365 "num_base_bdevs": 4, 00:10:19.365 "num_base_bdevs_discovered": 2, 00:10:19.365 "num_base_bdevs_operational": 4, 00:10:19.365 "base_bdevs_list": [ 00:10:19.365 { 00:10:19.365 "name": "BaseBdev1", 00:10:19.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.365 "is_configured": false, 00:10:19.365 "data_offset": 0, 00:10:19.365 "data_size": 0 00:10:19.365 }, 00:10:19.365 { 00:10:19.365 "name": null, 00:10:19.365 "uuid": "256ebd08-1af6-4530-a57a-b200efff626e", 00:10:19.365 "is_configured": false, 00:10:19.365 "data_offset": 0, 00:10:19.365 "data_size": 65536 00:10:19.365 }, 00:10:19.365 { 00:10:19.366 "name": "BaseBdev3", 00:10:19.366 "uuid": "3fa64264-5aba-4014-9c73-8c2581033a5c", 00:10:19.366 "is_configured": true, 00:10:19.366 "data_offset": 0, 00:10:19.366 "data_size": 65536 00:10:19.366 }, 00:10:19.366 { 00:10:19.366 "name": "BaseBdev4", 00:10:19.366 "uuid": "52635a23-96a4-4463-86a1-d4b1d12096ff", 00:10:19.366 "is_configured": true, 00:10:19.366 "data_offset": 0, 00:10:19.366 "data_size": 65536 00:10:19.366 } 00:10:19.366 ] 00:10:19.366 }' 00:10:19.366 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.366 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.626 [2024-12-07 05:37:52.951532] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:19.626 BaseBdev1 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.626 [ 00:10:19.626 { 00:10:19.626 "name": "BaseBdev1", 00:10:19.626 "aliases": [ 00:10:19.626 "f35645e2-219d-4ba9-9ae4-199d27188bb5" 00:10:19.626 ], 00:10:19.626 "product_name": "Malloc disk", 00:10:19.626 "block_size": 512, 00:10:19.626 "num_blocks": 65536, 00:10:19.626 "uuid": "f35645e2-219d-4ba9-9ae4-199d27188bb5", 00:10:19.626 "assigned_rate_limits": { 00:10:19.626 "rw_ios_per_sec": 0, 00:10:19.626 "rw_mbytes_per_sec": 0, 00:10:19.626 "r_mbytes_per_sec": 0, 00:10:19.626 "w_mbytes_per_sec": 0 00:10:19.626 }, 00:10:19.626 "claimed": true, 00:10:19.626 "claim_type": "exclusive_write", 00:10:19.626 "zoned": false, 00:10:19.626 "supported_io_types": { 00:10:19.626 "read": true, 00:10:19.626 "write": true, 00:10:19.626 "unmap": true, 00:10:19.626 "flush": true, 00:10:19.626 "reset": true, 00:10:19.626 "nvme_admin": false, 00:10:19.626 "nvme_io": false, 00:10:19.626 "nvme_io_md": false, 00:10:19.626 "write_zeroes": true, 00:10:19.626 "zcopy": true, 00:10:19.626 "get_zone_info": false, 00:10:19.626 "zone_management": false, 00:10:19.626 "zone_append": false, 00:10:19.626 "compare": false, 00:10:19.626 "compare_and_write": false, 00:10:19.626 "abort": true, 00:10:19.626 "seek_hole": false, 00:10:19.626 "seek_data": false, 00:10:19.626 "copy": true, 00:10:19.626 "nvme_iov_md": false 00:10:19.626 }, 00:10:19.626 "memory_domains": [ 00:10:19.626 { 00:10:19.626 "dma_device_id": "system", 00:10:19.626 "dma_device_type": 1 00:10:19.626 }, 00:10:19.626 { 00:10:19.626 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:19.626 "dma_device_type": 2 00:10:19.626 } 00:10:19.626 ], 00:10:19.626 "driver_specific": {} 00:10:19.626 } 00:10:19.626 ] 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.626 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.885 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.885 05:37:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:19.885 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.885 05:37:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.885 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.885 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.885 "name": "Existed_Raid", 00:10:19.885 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.885 "strip_size_kb": 0, 00:10:19.885 "state": "configuring", 00:10:19.885 "raid_level": "raid1", 00:10:19.885 "superblock": false, 00:10:19.885 "num_base_bdevs": 4, 00:10:19.885 "num_base_bdevs_discovered": 3, 00:10:19.885 "num_base_bdevs_operational": 4, 00:10:19.885 "base_bdevs_list": [ 00:10:19.885 { 00:10:19.885 "name": "BaseBdev1", 00:10:19.885 "uuid": "f35645e2-219d-4ba9-9ae4-199d27188bb5", 00:10:19.885 "is_configured": true, 00:10:19.885 "data_offset": 0, 00:10:19.885 "data_size": 65536 00:10:19.885 }, 00:10:19.885 { 00:10:19.885 "name": null, 00:10:19.885 "uuid": "256ebd08-1af6-4530-a57a-b200efff626e", 00:10:19.885 "is_configured": false, 00:10:19.885 "data_offset": 0, 00:10:19.885 "data_size": 65536 00:10:19.885 }, 00:10:19.885 { 00:10:19.885 "name": "BaseBdev3", 00:10:19.885 "uuid": "3fa64264-5aba-4014-9c73-8c2581033a5c", 00:10:19.885 "is_configured": true, 00:10:19.885 "data_offset": 0, 00:10:19.885 "data_size": 65536 00:10:19.885 }, 00:10:19.885 { 00:10:19.885 "name": "BaseBdev4", 00:10:19.885 "uuid": "52635a23-96a4-4463-86a1-d4b1d12096ff", 00:10:19.885 "is_configured": true, 00:10:19.885 "data_offset": 0, 00:10:19.885 "data_size": 65536 00:10:19.885 } 00:10:19.885 ] 00:10:19.885 }' 00:10:19.885 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.885 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.144 [2024-12-07 05:37:53.486673] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:20.144 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.145 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:20.404 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.404 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.404 "name": "Existed_Raid", 00:10:20.404 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.404 "strip_size_kb": 0, 00:10:20.404 "state": "configuring", 00:10:20.404 "raid_level": "raid1", 00:10:20.404 "superblock": false, 00:10:20.404 "num_base_bdevs": 4, 00:10:20.404 "num_base_bdevs_discovered": 2, 00:10:20.404 "num_base_bdevs_operational": 4, 00:10:20.404 "base_bdevs_list": [ 00:10:20.404 { 00:10:20.404 "name": "BaseBdev1", 00:10:20.404 "uuid": "f35645e2-219d-4ba9-9ae4-199d27188bb5", 00:10:20.404 "is_configured": true, 00:10:20.404 "data_offset": 0, 00:10:20.404 "data_size": 65536 00:10:20.404 }, 00:10:20.404 { 00:10:20.404 "name": null, 00:10:20.404 "uuid": "256ebd08-1af6-4530-a57a-b200efff626e", 00:10:20.404 "is_configured": false, 00:10:20.404 "data_offset": 0, 00:10:20.404 "data_size": 65536 00:10:20.404 }, 00:10:20.404 { 00:10:20.404 "name": null, 00:10:20.404 "uuid": "3fa64264-5aba-4014-9c73-8c2581033a5c", 00:10:20.404 "is_configured": false, 00:10:20.404 "data_offset": 0, 00:10:20.404 "data_size": 65536 00:10:20.404 }, 00:10:20.404 { 00:10:20.404 "name": "BaseBdev4", 00:10:20.404 "uuid": "52635a23-96a4-4463-86a1-d4b1d12096ff", 00:10:20.404 "is_configured": true, 00:10:20.404 "data_offset": 0, 00:10:20.404 "data_size": 65536 00:10:20.404 } 00:10:20.404 ] 00:10:20.404 }' 00:10:20.404 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.404 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.663 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.663 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.663 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.663 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:20.663 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.663 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:20.663 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.664 [2024-12-07 05:37:53.957946] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.664 05:37:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.664 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.664 "name": "Existed_Raid", 00:10:20.664 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.664 "strip_size_kb": 0, 00:10:20.664 "state": "configuring", 00:10:20.664 "raid_level": "raid1", 00:10:20.664 "superblock": false, 00:10:20.664 "num_base_bdevs": 4, 00:10:20.664 "num_base_bdevs_discovered": 3, 00:10:20.664 "num_base_bdevs_operational": 4, 00:10:20.664 "base_bdevs_list": [ 00:10:20.664 { 00:10:20.664 "name": "BaseBdev1", 00:10:20.664 "uuid": "f35645e2-219d-4ba9-9ae4-199d27188bb5", 00:10:20.664 "is_configured": true, 00:10:20.664 "data_offset": 0, 00:10:20.664 "data_size": 65536 00:10:20.664 }, 00:10:20.664 { 00:10:20.664 "name": null, 00:10:20.664 "uuid": "256ebd08-1af6-4530-a57a-b200efff626e", 00:10:20.664 "is_configured": false, 00:10:20.664 "data_offset": 0, 00:10:20.664 "data_size": 65536 00:10:20.664 }, 00:10:20.664 { 00:10:20.664 "name": "BaseBdev3", 00:10:20.664 "uuid": "3fa64264-5aba-4014-9c73-8c2581033a5c", 00:10:20.664 "is_configured": true, 00:10:20.664 "data_offset": 0, 00:10:20.664 "data_size": 65536 00:10:20.664 }, 00:10:20.664 { 00:10:20.664 "name": "BaseBdev4", 00:10:20.664 "uuid": "52635a23-96a4-4463-86a1-d4b1d12096ff", 00:10:20.664 "is_configured": true, 00:10:20.664 "data_offset": 0, 00:10:20.664 "data_size": 65536 00:10:20.664 } 00:10:20.664 ] 00:10:20.664 }' 00:10:20.664 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.664 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.233 [2024-12-07 05:37:54.413216] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:21.233 "name": "Existed_Raid", 00:10:21.233 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:21.233 "strip_size_kb": 0, 00:10:21.233 "state": "configuring", 00:10:21.233 "raid_level": "raid1", 00:10:21.233 "superblock": false, 00:10:21.233 "num_base_bdevs": 4, 00:10:21.233 "num_base_bdevs_discovered": 2, 00:10:21.233 "num_base_bdevs_operational": 4, 00:10:21.233 "base_bdevs_list": [ 00:10:21.233 { 00:10:21.233 "name": null, 00:10:21.233 "uuid": "f35645e2-219d-4ba9-9ae4-199d27188bb5", 00:10:21.233 "is_configured": false, 00:10:21.233 "data_offset": 0, 00:10:21.233 "data_size": 65536 00:10:21.233 }, 00:10:21.233 { 00:10:21.233 "name": null, 00:10:21.233 "uuid": "256ebd08-1af6-4530-a57a-b200efff626e", 00:10:21.233 "is_configured": false, 00:10:21.233 "data_offset": 0, 00:10:21.233 "data_size": 65536 00:10:21.233 }, 00:10:21.233 { 00:10:21.233 "name": "BaseBdev3", 00:10:21.233 "uuid": "3fa64264-5aba-4014-9c73-8c2581033a5c", 00:10:21.233 "is_configured": true, 00:10:21.233 "data_offset": 0, 00:10:21.233 "data_size": 65536 00:10:21.233 }, 00:10:21.233 { 00:10:21.233 "name": "BaseBdev4", 00:10:21.233 "uuid": "52635a23-96a4-4463-86a1-d4b1d12096ff", 00:10:21.233 "is_configured": true, 00:10:21.233 "data_offset": 0, 00:10:21.233 "data_size": 65536 00:10:21.233 } 00:10:21.233 ] 00:10:21.233 }' 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:21.233 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.492 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:21.492 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.492 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.492 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.751 [2024-12-07 05:37:54.880287] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.751 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:21.751 "name": "Existed_Raid", 00:10:21.751 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:21.751 "strip_size_kb": 0, 00:10:21.751 "state": "configuring", 00:10:21.751 "raid_level": "raid1", 00:10:21.751 "superblock": false, 00:10:21.751 "num_base_bdevs": 4, 00:10:21.751 "num_base_bdevs_discovered": 3, 00:10:21.751 "num_base_bdevs_operational": 4, 00:10:21.751 "base_bdevs_list": [ 00:10:21.751 { 00:10:21.751 "name": null, 00:10:21.751 "uuid": "f35645e2-219d-4ba9-9ae4-199d27188bb5", 00:10:21.751 "is_configured": false, 00:10:21.751 "data_offset": 0, 00:10:21.751 "data_size": 65536 00:10:21.751 }, 00:10:21.751 { 00:10:21.751 "name": "BaseBdev2", 00:10:21.751 "uuid": "256ebd08-1af6-4530-a57a-b200efff626e", 00:10:21.751 "is_configured": true, 00:10:21.751 "data_offset": 0, 00:10:21.751 "data_size": 65536 00:10:21.751 }, 00:10:21.751 { 00:10:21.751 "name": "BaseBdev3", 00:10:21.752 "uuid": "3fa64264-5aba-4014-9c73-8c2581033a5c", 00:10:21.752 "is_configured": true, 00:10:21.752 "data_offset": 0, 00:10:21.752 "data_size": 65536 00:10:21.752 }, 00:10:21.752 { 00:10:21.752 "name": "BaseBdev4", 00:10:21.752 "uuid": "52635a23-96a4-4463-86a1-d4b1d12096ff", 00:10:21.752 "is_configured": true, 00:10:21.752 "data_offset": 0, 00:10:21.752 "data_size": 65536 00:10:21.752 } 00:10:21.752 ] 00:10:21.752 }' 00:10:21.752 05:37:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:21.752 05:37:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.010 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:22.010 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.010 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.010 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.010 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.010 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:22.010 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.010 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.010 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:22.010 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.010 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u f35645e2-219d-4ba9-9ae4-199d27188bb5 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.270 [2024-12-07 05:37:55.416347] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:22.270 [2024-12-07 05:37:55.416407] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:22.270 [2024-12-07 05:37:55.416421] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:22.270 [2024-12-07 05:37:55.416714] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:22.270 [2024-12-07 05:37:55.416892] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:22.270 [2024-12-07 05:37:55.416902] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:22.270 [2024-12-07 05:37:55.417126] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:22.270 NewBaseBdev 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.270 [ 00:10:22.270 { 00:10:22.270 "name": "NewBaseBdev", 00:10:22.270 "aliases": [ 00:10:22.270 "f35645e2-219d-4ba9-9ae4-199d27188bb5" 00:10:22.270 ], 00:10:22.270 "product_name": "Malloc disk", 00:10:22.270 "block_size": 512, 00:10:22.270 "num_blocks": 65536, 00:10:22.270 "uuid": "f35645e2-219d-4ba9-9ae4-199d27188bb5", 00:10:22.270 "assigned_rate_limits": { 00:10:22.270 "rw_ios_per_sec": 0, 00:10:22.270 "rw_mbytes_per_sec": 0, 00:10:22.270 "r_mbytes_per_sec": 0, 00:10:22.270 "w_mbytes_per_sec": 0 00:10:22.270 }, 00:10:22.270 "claimed": true, 00:10:22.270 "claim_type": "exclusive_write", 00:10:22.270 "zoned": false, 00:10:22.270 "supported_io_types": { 00:10:22.270 "read": true, 00:10:22.270 "write": true, 00:10:22.270 "unmap": true, 00:10:22.270 "flush": true, 00:10:22.270 "reset": true, 00:10:22.270 "nvme_admin": false, 00:10:22.270 "nvme_io": false, 00:10:22.270 "nvme_io_md": false, 00:10:22.270 "write_zeroes": true, 00:10:22.270 "zcopy": true, 00:10:22.270 "get_zone_info": false, 00:10:22.270 "zone_management": false, 00:10:22.270 "zone_append": false, 00:10:22.270 "compare": false, 00:10:22.270 "compare_and_write": false, 00:10:22.270 "abort": true, 00:10:22.270 "seek_hole": false, 00:10:22.270 "seek_data": false, 00:10:22.270 "copy": true, 00:10:22.270 "nvme_iov_md": false 00:10:22.270 }, 00:10:22.270 "memory_domains": [ 00:10:22.270 { 00:10:22.270 "dma_device_id": "system", 00:10:22.270 "dma_device_type": 1 00:10:22.270 }, 00:10:22.270 { 00:10:22.270 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.270 "dma_device_type": 2 00:10:22.270 } 00:10:22.270 ], 00:10:22.270 "driver_specific": {} 00:10:22.270 } 00:10:22.270 ] 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.270 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:22.270 "name": "Existed_Raid", 00:10:22.270 "uuid": "4dc8463c-a01f-4ab1-98d0-fdc769d8bab9", 00:10:22.270 "strip_size_kb": 0, 00:10:22.270 "state": "online", 00:10:22.270 "raid_level": "raid1", 00:10:22.270 "superblock": false, 00:10:22.270 "num_base_bdevs": 4, 00:10:22.270 "num_base_bdevs_discovered": 4, 00:10:22.270 "num_base_bdevs_operational": 4, 00:10:22.270 "base_bdevs_list": [ 00:10:22.270 { 00:10:22.270 "name": "NewBaseBdev", 00:10:22.270 "uuid": "f35645e2-219d-4ba9-9ae4-199d27188bb5", 00:10:22.270 "is_configured": true, 00:10:22.270 "data_offset": 0, 00:10:22.270 "data_size": 65536 00:10:22.270 }, 00:10:22.270 { 00:10:22.270 "name": "BaseBdev2", 00:10:22.270 "uuid": "256ebd08-1af6-4530-a57a-b200efff626e", 00:10:22.270 "is_configured": true, 00:10:22.270 "data_offset": 0, 00:10:22.270 "data_size": 65536 00:10:22.270 }, 00:10:22.270 { 00:10:22.270 "name": "BaseBdev3", 00:10:22.270 "uuid": "3fa64264-5aba-4014-9c73-8c2581033a5c", 00:10:22.270 "is_configured": true, 00:10:22.270 "data_offset": 0, 00:10:22.270 "data_size": 65536 00:10:22.270 }, 00:10:22.270 { 00:10:22.270 "name": "BaseBdev4", 00:10:22.270 "uuid": "52635a23-96a4-4463-86a1-d4b1d12096ff", 00:10:22.270 "is_configured": true, 00:10:22.270 "data_offset": 0, 00:10:22.270 "data_size": 65536 00:10:22.271 } 00:10:22.271 ] 00:10:22.271 }' 00:10:22.271 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:22.271 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.530 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:22.530 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:22.530 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:22.530 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:22.530 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:22.530 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:22.530 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:22.530 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:22.530 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.530 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.530 [2024-12-07 05:37:55.887981] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:22.790 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.790 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:22.790 "name": "Existed_Raid", 00:10:22.790 "aliases": [ 00:10:22.790 "4dc8463c-a01f-4ab1-98d0-fdc769d8bab9" 00:10:22.790 ], 00:10:22.790 "product_name": "Raid Volume", 00:10:22.790 "block_size": 512, 00:10:22.790 "num_blocks": 65536, 00:10:22.790 "uuid": "4dc8463c-a01f-4ab1-98d0-fdc769d8bab9", 00:10:22.790 "assigned_rate_limits": { 00:10:22.790 "rw_ios_per_sec": 0, 00:10:22.790 "rw_mbytes_per_sec": 0, 00:10:22.790 "r_mbytes_per_sec": 0, 00:10:22.790 "w_mbytes_per_sec": 0 00:10:22.790 }, 00:10:22.790 "claimed": false, 00:10:22.790 "zoned": false, 00:10:22.790 "supported_io_types": { 00:10:22.790 "read": true, 00:10:22.790 "write": true, 00:10:22.790 "unmap": false, 00:10:22.790 "flush": false, 00:10:22.790 "reset": true, 00:10:22.790 "nvme_admin": false, 00:10:22.790 "nvme_io": false, 00:10:22.790 "nvme_io_md": false, 00:10:22.790 "write_zeroes": true, 00:10:22.790 "zcopy": false, 00:10:22.790 "get_zone_info": false, 00:10:22.790 "zone_management": false, 00:10:22.790 "zone_append": false, 00:10:22.790 "compare": false, 00:10:22.790 "compare_and_write": false, 00:10:22.790 "abort": false, 00:10:22.790 "seek_hole": false, 00:10:22.790 "seek_data": false, 00:10:22.790 "copy": false, 00:10:22.790 "nvme_iov_md": false 00:10:22.790 }, 00:10:22.790 "memory_domains": [ 00:10:22.790 { 00:10:22.790 "dma_device_id": "system", 00:10:22.790 "dma_device_type": 1 00:10:22.790 }, 00:10:22.790 { 00:10:22.790 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.790 "dma_device_type": 2 00:10:22.790 }, 00:10:22.790 { 00:10:22.790 "dma_device_id": "system", 00:10:22.790 "dma_device_type": 1 00:10:22.790 }, 00:10:22.790 { 00:10:22.790 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.790 "dma_device_type": 2 00:10:22.790 }, 00:10:22.790 { 00:10:22.790 "dma_device_id": "system", 00:10:22.790 "dma_device_type": 1 00:10:22.790 }, 00:10:22.790 { 00:10:22.790 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.790 "dma_device_type": 2 00:10:22.790 }, 00:10:22.790 { 00:10:22.790 "dma_device_id": "system", 00:10:22.790 "dma_device_type": 1 00:10:22.790 }, 00:10:22.790 { 00:10:22.790 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.790 "dma_device_type": 2 00:10:22.790 } 00:10:22.790 ], 00:10:22.790 "driver_specific": { 00:10:22.790 "raid": { 00:10:22.790 "uuid": "4dc8463c-a01f-4ab1-98d0-fdc769d8bab9", 00:10:22.790 "strip_size_kb": 0, 00:10:22.790 "state": "online", 00:10:22.790 "raid_level": "raid1", 00:10:22.790 "superblock": false, 00:10:22.790 "num_base_bdevs": 4, 00:10:22.790 "num_base_bdevs_discovered": 4, 00:10:22.790 "num_base_bdevs_operational": 4, 00:10:22.790 "base_bdevs_list": [ 00:10:22.790 { 00:10:22.790 "name": "NewBaseBdev", 00:10:22.790 "uuid": "f35645e2-219d-4ba9-9ae4-199d27188bb5", 00:10:22.790 "is_configured": true, 00:10:22.790 "data_offset": 0, 00:10:22.790 "data_size": 65536 00:10:22.790 }, 00:10:22.790 { 00:10:22.790 "name": "BaseBdev2", 00:10:22.790 "uuid": "256ebd08-1af6-4530-a57a-b200efff626e", 00:10:22.790 "is_configured": true, 00:10:22.790 "data_offset": 0, 00:10:22.790 "data_size": 65536 00:10:22.790 }, 00:10:22.790 { 00:10:22.790 "name": "BaseBdev3", 00:10:22.790 "uuid": "3fa64264-5aba-4014-9c73-8c2581033a5c", 00:10:22.790 "is_configured": true, 00:10:22.790 "data_offset": 0, 00:10:22.790 "data_size": 65536 00:10:22.790 }, 00:10:22.790 { 00:10:22.790 "name": "BaseBdev4", 00:10:22.790 "uuid": "52635a23-96a4-4463-86a1-d4b1d12096ff", 00:10:22.790 "is_configured": true, 00:10:22.790 "data_offset": 0, 00:10:22.790 "data_size": 65536 00:10:22.790 } 00:10:22.790 ] 00:10:22.790 } 00:10:22.790 } 00:10:22.790 }' 00:10:22.790 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:22.790 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:22.790 BaseBdev2 00:10:22.790 BaseBdev3 00:10:22.790 BaseBdev4' 00:10:22.790 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.790 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:22.790 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.790 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.790 05:37:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:22.790 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.790 05:37:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.790 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.790 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:22.790 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:22.790 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.790 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:22.790 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.790 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.790 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.790 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.791 [2024-12-07 05:37:56.143130] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:22.791 [2024-12-07 05:37:56.143173] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:22.791 [2024-12-07 05:37:56.143259] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:22.791 [2024-12-07 05:37:56.143550] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:22.791 [2024-12-07 05:37:56.143567] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 83629 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 83629 ']' 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 83629 00:10:22.791 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:10:23.050 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:23.050 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83629 00:10:23.050 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:23.050 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:23.050 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83629' 00:10:23.050 killing process with pid 83629 00:10:23.050 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 83629 00:10:23.050 [2024-12-07 05:37:56.191919] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:23.050 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 83629 00:10:23.050 [2024-12-07 05:37:56.271387] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:23.310 05:37:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:23.310 00:10:23.310 real 0m9.604s 00:10:23.310 user 0m16.288s 00:10:23.310 sys 0m1.937s 00:10:23.310 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:23.310 ************************************ 00:10:23.310 END TEST raid_state_function_test 00:10:23.310 ************************************ 00:10:23.310 05:37:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.310 05:37:56 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:10:23.310 05:37:56 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:23.310 05:37:56 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:23.310 05:37:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:23.310 ************************************ 00:10:23.310 START TEST raid_state_function_test_sb 00:10:23.310 ************************************ 00:10:23.310 05:37:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 true 00:10:23.310 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:23.310 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:23.310 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:23.310 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:23.570 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=84278 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84278' 00:10:23.571 Process raid pid: 84278 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 84278 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 84278 ']' 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:23.571 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:23.571 05:37:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.571 [2024-12-07 05:37:56.763375] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:10:23.571 [2024-12-07 05:37:56.763591] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:23.571 [2024-12-07 05:37:56.919688] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:23.830 [2024-12-07 05:37:56.963828] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:23.830 [2024-12-07 05:37:57.041877] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:23.830 [2024-12-07 05:37:57.042007] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.399 [2024-12-07 05:37:57.598407] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:24.399 [2024-12-07 05:37:57.598530] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:24.399 [2024-12-07 05:37:57.598555] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:24.399 [2024-12-07 05:37:57.598568] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:24.399 [2024-12-07 05:37:57.598574] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:24.399 [2024-12-07 05:37:57.598746] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:24.399 [2024-12-07 05:37:57.598753] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:24.399 [2024-12-07 05:37:57.598762] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:24.399 "name": "Existed_Raid", 00:10:24.399 "uuid": "6f17e738-33a6-4f12-b589-021fd4cd11ae", 00:10:24.399 "strip_size_kb": 0, 00:10:24.399 "state": "configuring", 00:10:24.399 "raid_level": "raid1", 00:10:24.399 "superblock": true, 00:10:24.399 "num_base_bdevs": 4, 00:10:24.399 "num_base_bdevs_discovered": 0, 00:10:24.399 "num_base_bdevs_operational": 4, 00:10:24.399 "base_bdevs_list": [ 00:10:24.399 { 00:10:24.399 "name": "BaseBdev1", 00:10:24.399 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.399 "is_configured": false, 00:10:24.399 "data_offset": 0, 00:10:24.399 "data_size": 0 00:10:24.399 }, 00:10:24.399 { 00:10:24.399 "name": "BaseBdev2", 00:10:24.399 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.399 "is_configured": false, 00:10:24.399 "data_offset": 0, 00:10:24.399 "data_size": 0 00:10:24.399 }, 00:10:24.399 { 00:10:24.399 "name": "BaseBdev3", 00:10:24.399 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.399 "is_configured": false, 00:10:24.399 "data_offset": 0, 00:10:24.399 "data_size": 0 00:10:24.399 }, 00:10:24.399 { 00:10:24.399 "name": "BaseBdev4", 00:10:24.399 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.399 "is_configured": false, 00:10:24.399 "data_offset": 0, 00:10:24.399 "data_size": 0 00:10:24.399 } 00:10:24.399 ] 00:10:24.399 }' 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:24.399 05:37:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.973 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:24.973 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.973 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.973 [2024-12-07 05:37:58.077585] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:24.973 [2024-12-07 05:37:58.077768] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:24.973 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.973 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:24.973 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.973 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.973 [2024-12-07 05:37:58.089565] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:24.973 [2024-12-07 05:37:58.089682] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:24.973 [2024-12-07 05:37:58.089715] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:24.973 [2024-12-07 05:37:58.089741] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:24.973 [2024-12-07 05:37:58.089759] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:24.973 [2024-12-07 05:37:58.089782] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:24.973 [2024-12-07 05:37:58.089800] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:24.973 [2024-12-07 05:37:58.089823] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.974 [2024-12-07 05:37:58.116672] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:24.974 BaseBdev1 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.974 [ 00:10:24.974 { 00:10:24.974 "name": "BaseBdev1", 00:10:24.974 "aliases": [ 00:10:24.974 "59fe56f0-ccc9-45d6-886b-c5863ad4110b" 00:10:24.974 ], 00:10:24.974 "product_name": "Malloc disk", 00:10:24.974 "block_size": 512, 00:10:24.974 "num_blocks": 65536, 00:10:24.974 "uuid": "59fe56f0-ccc9-45d6-886b-c5863ad4110b", 00:10:24.974 "assigned_rate_limits": { 00:10:24.974 "rw_ios_per_sec": 0, 00:10:24.974 "rw_mbytes_per_sec": 0, 00:10:24.974 "r_mbytes_per_sec": 0, 00:10:24.974 "w_mbytes_per_sec": 0 00:10:24.974 }, 00:10:24.974 "claimed": true, 00:10:24.974 "claim_type": "exclusive_write", 00:10:24.974 "zoned": false, 00:10:24.974 "supported_io_types": { 00:10:24.974 "read": true, 00:10:24.974 "write": true, 00:10:24.974 "unmap": true, 00:10:24.974 "flush": true, 00:10:24.974 "reset": true, 00:10:24.974 "nvme_admin": false, 00:10:24.974 "nvme_io": false, 00:10:24.974 "nvme_io_md": false, 00:10:24.974 "write_zeroes": true, 00:10:24.974 "zcopy": true, 00:10:24.974 "get_zone_info": false, 00:10:24.974 "zone_management": false, 00:10:24.974 "zone_append": false, 00:10:24.974 "compare": false, 00:10:24.974 "compare_and_write": false, 00:10:24.974 "abort": true, 00:10:24.974 "seek_hole": false, 00:10:24.974 "seek_data": false, 00:10:24.974 "copy": true, 00:10:24.974 "nvme_iov_md": false 00:10:24.974 }, 00:10:24.974 "memory_domains": [ 00:10:24.974 { 00:10:24.974 "dma_device_id": "system", 00:10:24.974 "dma_device_type": 1 00:10:24.974 }, 00:10:24.974 { 00:10:24.974 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:24.974 "dma_device_type": 2 00:10:24.974 } 00:10:24.974 ], 00:10:24.974 "driver_specific": {} 00:10:24.974 } 00:10:24.974 ] 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:24.974 "name": "Existed_Raid", 00:10:24.974 "uuid": "c72defd0-8584-45a9-9d40-d20a0e96d574", 00:10:24.974 "strip_size_kb": 0, 00:10:24.974 "state": "configuring", 00:10:24.974 "raid_level": "raid1", 00:10:24.974 "superblock": true, 00:10:24.974 "num_base_bdevs": 4, 00:10:24.974 "num_base_bdevs_discovered": 1, 00:10:24.974 "num_base_bdevs_operational": 4, 00:10:24.974 "base_bdevs_list": [ 00:10:24.974 { 00:10:24.974 "name": "BaseBdev1", 00:10:24.974 "uuid": "59fe56f0-ccc9-45d6-886b-c5863ad4110b", 00:10:24.974 "is_configured": true, 00:10:24.974 "data_offset": 2048, 00:10:24.974 "data_size": 63488 00:10:24.974 }, 00:10:24.974 { 00:10:24.974 "name": "BaseBdev2", 00:10:24.974 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.974 "is_configured": false, 00:10:24.974 "data_offset": 0, 00:10:24.974 "data_size": 0 00:10:24.974 }, 00:10:24.974 { 00:10:24.974 "name": "BaseBdev3", 00:10:24.974 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.974 "is_configured": false, 00:10:24.974 "data_offset": 0, 00:10:24.974 "data_size": 0 00:10:24.974 }, 00:10:24.974 { 00:10:24.974 "name": "BaseBdev4", 00:10:24.974 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.974 "is_configured": false, 00:10:24.974 "data_offset": 0, 00:10:24.974 "data_size": 0 00:10:24.974 } 00:10:24.974 ] 00:10:24.974 }' 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:24.974 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.544 [2024-12-07 05:37:58.619881] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:25.544 [2024-12-07 05:37:58.620039] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.544 [2024-12-07 05:37:58.631883] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:25.544 [2024-12-07 05:37:58.634107] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:25.544 [2024-12-07 05:37:58.634187] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:25.544 [2024-12-07 05:37:58.634221] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:25.544 [2024-12-07 05:37:58.634244] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:25.544 [2024-12-07 05:37:58.634288] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:25.544 [2024-12-07 05:37:58.634310] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.544 "name": "Existed_Raid", 00:10:25.544 "uuid": "1c5e9a9e-dcbf-45b8-b679-4c6e8cb5415f", 00:10:25.544 "strip_size_kb": 0, 00:10:25.544 "state": "configuring", 00:10:25.544 "raid_level": "raid1", 00:10:25.544 "superblock": true, 00:10:25.544 "num_base_bdevs": 4, 00:10:25.544 "num_base_bdevs_discovered": 1, 00:10:25.544 "num_base_bdevs_operational": 4, 00:10:25.544 "base_bdevs_list": [ 00:10:25.544 { 00:10:25.544 "name": "BaseBdev1", 00:10:25.544 "uuid": "59fe56f0-ccc9-45d6-886b-c5863ad4110b", 00:10:25.544 "is_configured": true, 00:10:25.544 "data_offset": 2048, 00:10:25.544 "data_size": 63488 00:10:25.544 }, 00:10:25.544 { 00:10:25.544 "name": "BaseBdev2", 00:10:25.544 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.544 "is_configured": false, 00:10:25.544 "data_offset": 0, 00:10:25.544 "data_size": 0 00:10:25.544 }, 00:10:25.544 { 00:10:25.544 "name": "BaseBdev3", 00:10:25.544 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.544 "is_configured": false, 00:10:25.544 "data_offset": 0, 00:10:25.544 "data_size": 0 00:10:25.544 }, 00:10:25.544 { 00:10:25.544 "name": "BaseBdev4", 00:10:25.544 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.544 "is_configured": false, 00:10:25.544 "data_offset": 0, 00:10:25.544 "data_size": 0 00:10:25.544 } 00:10:25.544 ] 00:10:25.544 }' 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.544 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.804 05:37:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:25.804 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.804 05:37:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.804 [2024-12-07 05:37:59.016149] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:25.804 BaseBdev2 00:10:25.804 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.804 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:25.804 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:25.804 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:25.804 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:25.804 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:25.804 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:25.804 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.805 [ 00:10:25.805 { 00:10:25.805 "name": "BaseBdev2", 00:10:25.805 "aliases": [ 00:10:25.805 "64e467e6-79ac-4865-8ed9-0b50d163f713" 00:10:25.805 ], 00:10:25.805 "product_name": "Malloc disk", 00:10:25.805 "block_size": 512, 00:10:25.805 "num_blocks": 65536, 00:10:25.805 "uuid": "64e467e6-79ac-4865-8ed9-0b50d163f713", 00:10:25.805 "assigned_rate_limits": { 00:10:25.805 "rw_ios_per_sec": 0, 00:10:25.805 "rw_mbytes_per_sec": 0, 00:10:25.805 "r_mbytes_per_sec": 0, 00:10:25.805 "w_mbytes_per_sec": 0 00:10:25.805 }, 00:10:25.805 "claimed": true, 00:10:25.805 "claim_type": "exclusive_write", 00:10:25.805 "zoned": false, 00:10:25.805 "supported_io_types": { 00:10:25.805 "read": true, 00:10:25.805 "write": true, 00:10:25.805 "unmap": true, 00:10:25.805 "flush": true, 00:10:25.805 "reset": true, 00:10:25.805 "nvme_admin": false, 00:10:25.805 "nvme_io": false, 00:10:25.805 "nvme_io_md": false, 00:10:25.805 "write_zeroes": true, 00:10:25.805 "zcopy": true, 00:10:25.805 "get_zone_info": false, 00:10:25.805 "zone_management": false, 00:10:25.805 "zone_append": false, 00:10:25.805 "compare": false, 00:10:25.805 "compare_and_write": false, 00:10:25.805 "abort": true, 00:10:25.805 "seek_hole": false, 00:10:25.805 "seek_data": false, 00:10:25.805 "copy": true, 00:10:25.805 "nvme_iov_md": false 00:10:25.805 }, 00:10:25.805 "memory_domains": [ 00:10:25.805 { 00:10:25.805 "dma_device_id": "system", 00:10:25.805 "dma_device_type": 1 00:10:25.805 }, 00:10:25.805 { 00:10:25.805 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.805 "dma_device_type": 2 00:10:25.805 } 00:10:25.805 ], 00:10:25.805 "driver_specific": {} 00:10:25.805 } 00:10:25.805 ] 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.805 "name": "Existed_Raid", 00:10:25.805 "uuid": "1c5e9a9e-dcbf-45b8-b679-4c6e8cb5415f", 00:10:25.805 "strip_size_kb": 0, 00:10:25.805 "state": "configuring", 00:10:25.805 "raid_level": "raid1", 00:10:25.805 "superblock": true, 00:10:25.805 "num_base_bdevs": 4, 00:10:25.805 "num_base_bdevs_discovered": 2, 00:10:25.805 "num_base_bdevs_operational": 4, 00:10:25.805 "base_bdevs_list": [ 00:10:25.805 { 00:10:25.805 "name": "BaseBdev1", 00:10:25.805 "uuid": "59fe56f0-ccc9-45d6-886b-c5863ad4110b", 00:10:25.805 "is_configured": true, 00:10:25.805 "data_offset": 2048, 00:10:25.805 "data_size": 63488 00:10:25.805 }, 00:10:25.805 { 00:10:25.805 "name": "BaseBdev2", 00:10:25.805 "uuid": "64e467e6-79ac-4865-8ed9-0b50d163f713", 00:10:25.805 "is_configured": true, 00:10:25.805 "data_offset": 2048, 00:10:25.805 "data_size": 63488 00:10:25.805 }, 00:10:25.805 { 00:10:25.805 "name": "BaseBdev3", 00:10:25.805 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.805 "is_configured": false, 00:10:25.805 "data_offset": 0, 00:10:25.805 "data_size": 0 00:10:25.805 }, 00:10:25.805 { 00:10:25.805 "name": "BaseBdev4", 00:10:25.805 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.805 "is_configured": false, 00:10:25.805 "data_offset": 0, 00:10:25.805 "data_size": 0 00:10:25.805 } 00:10:25.805 ] 00:10:25.805 }' 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.805 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.375 [2024-12-07 05:37:59.535329] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:26.375 BaseBdev3 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.375 [ 00:10:26.375 { 00:10:26.375 "name": "BaseBdev3", 00:10:26.375 "aliases": [ 00:10:26.375 "d39cc5c0-a976-4dbf-a4d4-9791b1f8ef3b" 00:10:26.375 ], 00:10:26.375 "product_name": "Malloc disk", 00:10:26.375 "block_size": 512, 00:10:26.375 "num_blocks": 65536, 00:10:26.375 "uuid": "d39cc5c0-a976-4dbf-a4d4-9791b1f8ef3b", 00:10:26.375 "assigned_rate_limits": { 00:10:26.375 "rw_ios_per_sec": 0, 00:10:26.375 "rw_mbytes_per_sec": 0, 00:10:26.375 "r_mbytes_per_sec": 0, 00:10:26.375 "w_mbytes_per_sec": 0 00:10:26.375 }, 00:10:26.375 "claimed": true, 00:10:26.375 "claim_type": "exclusive_write", 00:10:26.375 "zoned": false, 00:10:26.375 "supported_io_types": { 00:10:26.375 "read": true, 00:10:26.375 "write": true, 00:10:26.375 "unmap": true, 00:10:26.375 "flush": true, 00:10:26.375 "reset": true, 00:10:26.375 "nvme_admin": false, 00:10:26.375 "nvme_io": false, 00:10:26.375 "nvme_io_md": false, 00:10:26.375 "write_zeroes": true, 00:10:26.375 "zcopy": true, 00:10:26.375 "get_zone_info": false, 00:10:26.375 "zone_management": false, 00:10:26.375 "zone_append": false, 00:10:26.375 "compare": false, 00:10:26.375 "compare_and_write": false, 00:10:26.375 "abort": true, 00:10:26.375 "seek_hole": false, 00:10:26.375 "seek_data": false, 00:10:26.375 "copy": true, 00:10:26.375 "nvme_iov_md": false 00:10:26.375 }, 00:10:26.375 "memory_domains": [ 00:10:26.375 { 00:10:26.375 "dma_device_id": "system", 00:10:26.375 "dma_device_type": 1 00:10:26.375 }, 00:10:26.375 { 00:10:26.375 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.375 "dma_device_type": 2 00:10:26.375 } 00:10:26.375 ], 00:10:26.375 "driver_specific": {} 00:10:26.375 } 00:10:26.375 ] 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:26.375 "name": "Existed_Raid", 00:10:26.375 "uuid": "1c5e9a9e-dcbf-45b8-b679-4c6e8cb5415f", 00:10:26.375 "strip_size_kb": 0, 00:10:26.375 "state": "configuring", 00:10:26.375 "raid_level": "raid1", 00:10:26.375 "superblock": true, 00:10:26.375 "num_base_bdevs": 4, 00:10:26.375 "num_base_bdevs_discovered": 3, 00:10:26.375 "num_base_bdevs_operational": 4, 00:10:26.375 "base_bdevs_list": [ 00:10:26.375 { 00:10:26.375 "name": "BaseBdev1", 00:10:26.375 "uuid": "59fe56f0-ccc9-45d6-886b-c5863ad4110b", 00:10:26.375 "is_configured": true, 00:10:26.375 "data_offset": 2048, 00:10:26.375 "data_size": 63488 00:10:26.375 }, 00:10:26.375 { 00:10:26.375 "name": "BaseBdev2", 00:10:26.375 "uuid": "64e467e6-79ac-4865-8ed9-0b50d163f713", 00:10:26.375 "is_configured": true, 00:10:26.375 "data_offset": 2048, 00:10:26.375 "data_size": 63488 00:10:26.375 }, 00:10:26.375 { 00:10:26.375 "name": "BaseBdev3", 00:10:26.375 "uuid": "d39cc5c0-a976-4dbf-a4d4-9791b1f8ef3b", 00:10:26.375 "is_configured": true, 00:10:26.375 "data_offset": 2048, 00:10:26.375 "data_size": 63488 00:10:26.375 }, 00:10:26.375 { 00:10:26.375 "name": "BaseBdev4", 00:10:26.375 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:26.375 "is_configured": false, 00:10:26.375 "data_offset": 0, 00:10:26.375 "data_size": 0 00:10:26.375 } 00:10:26.375 ] 00:10:26.375 }' 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:26.375 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.635 05:37:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:26.635 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.635 05:37:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.895 [2024-12-07 05:38:00.003511] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:26.895 [2024-12-07 05:38:00.003784] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:26.895 [2024-12-07 05:38:00.003807] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:26.895 BaseBdev4 00:10:26.895 [2024-12-07 05:38:00.004104] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:26.895 [2024-12-07 05:38:00.004289] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:26.895 [2024-12-07 05:38:00.004303] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:26.895 [2024-12-07 05:38:00.004455] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.895 [ 00:10:26.895 { 00:10:26.895 "name": "BaseBdev4", 00:10:26.895 "aliases": [ 00:10:26.895 "d5893986-5aaa-4004-be6d-d519a5bbc3f5" 00:10:26.895 ], 00:10:26.895 "product_name": "Malloc disk", 00:10:26.895 "block_size": 512, 00:10:26.895 "num_blocks": 65536, 00:10:26.895 "uuid": "d5893986-5aaa-4004-be6d-d519a5bbc3f5", 00:10:26.895 "assigned_rate_limits": { 00:10:26.895 "rw_ios_per_sec": 0, 00:10:26.895 "rw_mbytes_per_sec": 0, 00:10:26.895 "r_mbytes_per_sec": 0, 00:10:26.895 "w_mbytes_per_sec": 0 00:10:26.895 }, 00:10:26.895 "claimed": true, 00:10:26.895 "claim_type": "exclusive_write", 00:10:26.895 "zoned": false, 00:10:26.895 "supported_io_types": { 00:10:26.895 "read": true, 00:10:26.895 "write": true, 00:10:26.895 "unmap": true, 00:10:26.895 "flush": true, 00:10:26.895 "reset": true, 00:10:26.895 "nvme_admin": false, 00:10:26.895 "nvme_io": false, 00:10:26.895 "nvme_io_md": false, 00:10:26.895 "write_zeroes": true, 00:10:26.895 "zcopy": true, 00:10:26.895 "get_zone_info": false, 00:10:26.895 "zone_management": false, 00:10:26.895 "zone_append": false, 00:10:26.895 "compare": false, 00:10:26.895 "compare_and_write": false, 00:10:26.895 "abort": true, 00:10:26.895 "seek_hole": false, 00:10:26.895 "seek_data": false, 00:10:26.895 "copy": true, 00:10:26.895 "nvme_iov_md": false 00:10:26.895 }, 00:10:26.895 "memory_domains": [ 00:10:26.895 { 00:10:26.895 "dma_device_id": "system", 00:10:26.895 "dma_device_type": 1 00:10:26.895 }, 00:10:26.895 { 00:10:26.895 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.895 "dma_device_type": 2 00:10:26.895 } 00:10:26.895 ], 00:10:26.895 "driver_specific": {} 00:10:26.895 } 00:10:26.895 ] 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:26.895 "name": "Existed_Raid", 00:10:26.895 "uuid": "1c5e9a9e-dcbf-45b8-b679-4c6e8cb5415f", 00:10:26.895 "strip_size_kb": 0, 00:10:26.895 "state": "online", 00:10:26.895 "raid_level": "raid1", 00:10:26.895 "superblock": true, 00:10:26.895 "num_base_bdevs": 4, 00:10:26.895 "num_base_bdevs_discovered": 4, 00:10:26.895 "num_base_bdevs_operational": 4, 00:10:26.895 "base_bdevs_list": [ 00:10:26.895 { 00:10:26.895 "name": "BaseBdev1", 00:10:26.895 "uuid": "59fe56f0-ccc9-45d6-886b-c5863ad4110b", 00:10:26.895 "is_configured": true, 00:10:26.895 "data_offset": 2048, 00:10:26.895 "data_size": 63488 00:10:26.895 }, 00:10:26.895 { 00:10:26.895 "name": "BaseBdev2", 00:10:26.895 "uuid": "64e467e6-79ac-4865-8ed9-0b50d163f713", 00:10:26.895 "is_configured": true, 00:10:26.895 "data_offset": 2048, 00:10:26.895 "data_size": 63488 00:10:26.895 }, 00:10:26.895 { 00:10:26.895 "name": "BaseBdev3", 00:10:26.895 "uuid": "d39cc5c0-a976-4dbf-a4d4-9791b1f8ef3b", 00:10:26.895 "is_configured": true, 00:10:26.895 "data_offset": 2048, 00:10:26.895 "data_size": 63488 00:10:26.895 }, 00:10:26.895 { 00:10:26.895 "name": "BaseBdev4", 00:10:26.895 "uuid": "d5893986-5aaa-4004-be6d-d519a5bbc3f5", 00:10:26.895 "is_configured": true, 00:10:26.895 "data_offset": 2048, 00:10:26.895 "data_size": 63488 00:10:26.895 } 00:10:26.895 ] 00:10:26.895 }' 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:26.895 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.155 [2024-12-07 05:38:00.467106] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.155 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:27.155 "name": "Existed_Raid", 00:10:27.155 "aliases": [ 00:10:27.155 "1c5e9a9e-dcbf-45b8-b679-4c6e8cb5415f" 00:10:27.155 ], 00:10:27.155 "product_name": "Raid Volume", 00:10:27.155 "block_size": 512, 00:10:27.155 "num_blocks": 63488, 00:10:27.155 "uuid": "1c5e9a9e-dcbf-45b8-b679-4c6e8cb5415f", 00:10:27.155 "assigned_rate_limits": { 00:10:27.155 "rw_ios_per_sec": 0, 00:10:27.155 "rw_mbytes_per_sec": 0, 00:10:27.155 "r_mbytes_per_sec": 0, 00:10:27.155 "w_mbytes_per_sec": 0 00:10:27.155 }, 00:10:27.155 "claimed": false, 00:10:27.155 "zoned": false, 00:10:27.155 "supported_io_types": { 00:10:27.155 "read": true, 00:10:27.155 "write": true, 00:10:27.155 "unmap": false, 00:10:27.155 "flush": false, 00:10:27.155 "reset": true, 00:10:27.155 "nvme_admin": false, 00:10:27.155 "nvme_io": false, 00:10:27.155 "nvme_io_md": false, 00:10:27.155 "write_zeroes": true, 00:10:27.155 "zcopy": false, 00:10:27.155 "get_zone_info": false, 00:10:27.155 "zone_management": false, 00:10:27.155 "zone_append": false, 00:10:27.155 "compare": false, 00:10:27.155 "compare_and_write": false, 00:10:27.155 "abort": false, 00:10:27.155 "seek_hole": false, 00:10:27.155 "seek_data": false, 00:10:27.155 "copy": false, 00:10:27.155 "nvme_iov_md": false 00:10:27.155 }, 00:10:27.155 "memory_domains": [ 00:10:27.155 { 00:10:27.155 "dma_device_id": "system", 00:10:27.155 "dma_device_type": 1 00:10:27.155 }, 00:10:27.155 { 00:10:27.155 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:27.155 "dma_device_type": 2 00:10:27.155 }, 00:10:27.155 { 00:10:27.155 "dma_device_id": "system", 00:10:27.155 "dma_device_type": 1 00:10:27.155 }, 00:10:27.155 { 00:10:27.155 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:27.155 "dma_device_type": 2 00:10:27.155 }, 00:10:27.155 { 00:10:27.155 "dma_device_id": "system", 00:10:27.155 "dma_device_type": 1 00:10:27.155 }, 00:10:27.155 { 00:10:27.155 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:27.155 "dma_device_type": 2 00:10:27.155 }, 00:10:27.155 { 00:10:27.155 "dma_device_id": "system", 00:10:27.155 "dma_device_type": 1 00:10:27.155 }, 00:10:27.155 { 00:10:27.155 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:27.155 "dma_device_type": 2 00:10:27.155 } 00:10:27.155 ], 00:10:27.155 "driver_specific": { 00:10:27.155 "raid": { 00:10:27.155 "uuid": "1c5e9a9e-dcbf-45b8-b679-4c6e8cb5415f", 00:10:27.155 "strip_size_kb": 0, 00:10:27.155 "state": "online", 00:10:27.155 "raid_level": "raid1", 00:10:27.155 "superblock": true, 00:10:27.155 "num_base_bdevs": 4, 00:10:27.155 "num_base_bdevs_discovered": 4, 00:10:27.155 "num_base_bdevs_operational": 4, 00:10:27.155 "base_bdevs_list": [ 00:10:27.155 { 00:10:27.155 "name": "BaseBdev1", 00:10:27.155 "uuid": "59fe56f0-ccc9-45d6-886b-c5863ad4110b", 00:10:27.155 "is_configured": true, 00:10:27.155 "data_offset": 2048, 00:10:27.155 "data_size": 63488 00:10:27.155 }, 00:10:27.155 { 00:10:27.155 "name": "BaseBdev2", 00:10:27.155 "uuid": "64e467e6-79ac-4865-8ed9-0b50d163f713", 00:10:27.155 "is_configured": true, 00:10:27.155 "data_offset": 2048, 00:10:27.155 "data_size": 63488 00:10:27.155 }, 00:10:27.155 { 00:10:27.155 "name": "BaseBdev3", 00:10:27.155 "uuid": "d39cc5c0-a976-4dbf-a4d4-9791b1f8ef3b", 00:10:27.155 "is_configured": true, 00:10:27.155 "data_offset": 2048, 00:10:27.155 "data_size": 63488 00:10:27.155 }, 00:10:27.155 { 00:10:27.155 "name": "BaseBdev4", 00:10:27.155 "uuid": "d5893986-5aaa-4004-be6d-d519a5bbc3f5", 00:10:27.155 "is_configured": true, 00:10:27.155 "data_offset": 2048, 00:10:27.156 "data_size": 63488 00:10:27.156 } 00:10:27.156 ] 00:10:27.156 } 00:10:27.156 } 00:10:27.156 }' 00:10:27.156 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:27.415 BaseBdev2 00:10:27.415 BaseBdev3 00:10:27.415 BaseBdev4' 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:27.415 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.416 [2024-12-07 05:38:00.706516] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.416 "name": "Existed_Raid", 00:10:27.416 "uuid": "1c5e9a9e-dcbf-45b8-b679-4c6e8cb5415f", 00:10:27.416 "strip_size_kb": 0, 00:10:27.416 "state": "online", 00:10:27.416 "raid_level": "raid1", 00:10:27.416 "superblock": true, 00:10:27.416 "num_base_bdevs": 4, 00:10:27.416 "num_base_bdevs_discovered": 3, 00:10:27.416 "num_base_bdevs_operational": 3, 00:10:27.416 "base_bdevs_list": [ 00:10:27.416 { 00:10:27.416 "name": null, 00:10:27.416 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.416 "is_configured": false, 00:10:27.416 "data_offset": 0, 00:10:27.416 "data_size": 63488 00:10:27.416 }, 00:10:27.416 { 00:10:27.416 "name": "BaseBdev2", 00:10:27.416 "uuid": "64e467e6-79ac-4865-8ed9-0b50d163f713", 00:10:27.416 "is_configured": true, 00:10:27.416 "data_offset": 2048, 00:10:27.416 "data_size": 63488 00:10:27.416 }, 00:10:27.416 { 00:10:27.416 "name": "BaseBdev3", 00:10:27.416 "uuid": "d39cc5c0-a976-4dbf-a4d4-9791b1f8ef3b", 00:10:27.416 "is_configured": true, 00:10:27.416 "data_offset": 2048, 00:10:27.416 "data_size": 63488 00:10:27.416 }, 00:10:27.416 { 00:10:27.416 "name": "BaseBdev4", 00:10:27.416 "uuid": "d5893986-5aaa-4004-be6d-d519a5bbc3f5", 00:10:27.416 "is_configured": true, 00:10:27.416 "data_offset": 2048, 00:10:27.416 "data_size": 63488 00:10:27.416 } 00:10:27.416 ] 00:10:27.416 }' 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.416 05:38:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.985 [2024-12-07 05:38:01.178584] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.985 [2024-12-07 05:38:01.259083] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.985 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.986 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.986 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.986 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:27.986 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:27.986 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:27.986 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.986 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.986 [2024-12-07 05:38:01.327694] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:27.986 [2024-12-07 05:38:01.327841] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:27.986 [2024-12-07 05:38:01.349112] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:27.986 [2024-12-07 05:38:01.349175] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:27.986 [2024-12-07 05:38:01.349191] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:27.986 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.986 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:27.986 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.246 BaseBdev2 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.246 [ 00:10:28.246 { 00:10:28.246 "name": "BaseBdev2", 00:10:28.246 "aliases": [ 00:10:28.246 "1a75b823-bf80-4661-8b93-e11702fdd45c" 00:10:28.246 ], 00:10:28.246 "product_name": "Malloc disk", 00:10:28.246 "block_size": 512, 00:10:28.246 "num_blocks": 65536, 00:10:28.246 "uuid": "1a75b823-bf80-4661-8b93-e11702fdd45c", 00:10:28.246 "assigned_rate_limits": { 00:10:28.246 "rw_ios_per_sec": 0, 00:10:28.246 "rw_mbytes_per_sec": 0, 00:10:28.246 "r_mbytes_per_sec": 0, 00:10:28.246 "w_mbytes_per_sec": 0 00:10:28.246 }, 00:10:28.246 "claimed": false, 00:10:28.246 "zoned": false, 00:10:28.246 "supported_io_types": { 00:10:28.246 "read": true, 00:10:28.246 "write": true, 00:10:28.246 "unmap": true, 00:10:28.246 "flush": true, 00:10:28.246 "reset": true, 00:10:28.246 "nvme_admin": false, 00:10:28.246 "nvme_io": false, 00:10:28.246 "nvme_io_md": false, 00:10:28.246 "write_zeroes": true, 00:10:28.246 "zcopy": true, 00:10:28.246 "get_zone_info": false, 00:10:28.246 "zone_management": false, 00:10:28.246 "zone_append": false, 00:10:28.246 "compare": false, 00:10:28.246 "compare_and_write": false, 00:10:28.246 "abort": true, 00:10:28.246 "seek_hole": false, 00:10:28.246 "seek_data": false, 00:10:28.246 "copy": true, 00:10:28.246 "nvme_iov_md": false 00:10:28.246 }, 00:10:28.246 "memory_domains": [ 00:10:28.246 { 00:10:28.246 "dma_device_id": "system", 00:10:28.246 "dma_device_type": 1 00:10:28.246 }, 00:10:28.246 { 00:10:28.246 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:28.246 "dma_device_type": 2 00:10:28.246 } 00:10:28.246 ], 00:10:28.246 "driver_specific": {} 00:10:28.246 } 00:10:28.246 ] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.246 BaseBdev3 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.246 [ 00:10:28.246 { 00:10:28.246 "name": "BaseBdev3", 00:10:28.246 "aliases": [ 00:10:28.246 "d7f13a84-a8af-490d-b5cb-89370d72595a" 00:10:28.246 ], 00:10:28.246 "product_name": "Malloc disk", 00:10:28.246 "block_size": 512, 00:10:28.246 "num_blocks": 65536, 00:10:28.246 "uuid": "d7f13a84-a8af-490d-b5cb-89370d72595a", 00:10:28.246 "assigned_rate_limits": { 00:10:28.246 "rw_ios_per_sec": 0, 00:10:28.246 "rw_mbytes_per_sec": 0, 00:10:28.246 "r_mbytes_per_sec": 0, 00:10:28.246 "w_mbytes_per_sec": 0 00:10:28.246 }, 00:10:28.246 "claimed": false, 00:10:28.246 "zoned": false, 00:10:28.246 "supported_io_types": { 00:10:28.246 "read": true, 00:10:28.246 "write": true, 00:10:28.246 "unmap": true, 00:10:28.246 "flush": true, 00:10:28.246 "reset": true, 00:10:28.246 "nvme_admin": false, 00:10:28.246 "nvme_io": false, 00:10:28.246 "nvme_io_md": false, 00:10:28.246 "write_zeroes": true, 00:10:28.246 "zcopy": true, 00:10:28.246 "get_zone_info": false, 00:10:28.246 "zone_management": false, 00:10:28.246 "zone_append": false, 00:10:28.246 "compare": false, 00:10:28.246 "compare_and_write": false, 00:10:28.246 "abort": true, 00:10:28.246 "seek_hole": false, 00:10:28.246 "seek_data": false, 00:10:28.246 "copy": true, 00:10:28.246 "nvme_iov_md": false 00:10:28.246 }, 00:10:28.246 "memory_domains": [ 00:10:28.246 { 00:10:28.246 "dma_device_id": "system", 00:10:28.246 "dma_device_type": 1 00:10:28.246 }, 00:10:28.246 { 00:10:28.246 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:28.246 "dma_device_type": 2 00:10:28.246 } 00:10:28.246 ], 00:10:28.246 "driver_specific": {} 00:10:28.246 } 00:10:28.246 ] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.246 BaseBdev4 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.246 [ 00:10:28.246 { 00:10:28.246 "name": "BaseBdev4", 00:10:28.246 "aliases": [ 00:10:28.246 "2d15a1b8-b660-422d-890f-ff60dbaf8961" 00:10:28.246 ], 00:10:28.246 "product_name": "Malloc disk", 00:10:28.246 "block_size": 512, 00:10:28.246 "num_blocks": 65536, 00:10:28.246 "uuid": "2d15a1b8-b660-422d-890f-ff60dbaf8961", 00:10:28.246 "assigned_rate_limits": { 00:10:28.246 "rw_ios_per_sec": 0, 00:10:28.246 "rw_mbytes_per_sec": 0, 00:10:28.246 "r_mbytes_per_sec": 0, 00:10:28.246 "w_mbytes_per_sec": 0 00:10:28.246 }, 00:10:28.246 "claimed": false, 00:10:28.246 "zoned": false, 00:10:28.246 "supported_io_types": { 00:10:28.246 "read": true, 00:10:28.246 "write": true, 00:10:28.246 "unmap": true, 00:10:28.246 "flush": true, 00:10:28.246 "reset": true, 00:10:28.246 "nvme_admin": false, 00:10:28.246 "nvme_io": false, 00:10:28.246 "nvme_io_md": false, 00:10:28.246 "write_zeroes": true, 00:10:28.246 "zcopy": true, 00:10:28.246 "get_zone_info": false, 00:10:28.246 "zone_management": false, 00:10:28.246 "zone_append": false, 00:10:28.246 "compare": false, 00:10:28.246 "compare_and_write": false, 00:10:28.246 "abort": true, 00:10:28.246 "seek_hole": false, 00:10:28.246 "seek_data": false, 00:10:28.246 "copy": true, 00:10:28.246 "nvme_iov_md": false 00:10:28.246 }, 00:10:28.246 "memory_domains": [ 00:10:28.246 { 00:10:28.246 "dma_device_id": "system", 00:10:28.246 "dma_device_type": 1 00:10:28.246 }, 00:10:28.246 { 00:10:28.246 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:28.246 "dma_device_type": 2 00:10:28.246 } 00:10:28.246 ], 00:10:28.246 "driver_specific": {} 00:10:28.246 } 00:10:28.246 ] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:28.246 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.247 [2024-12-07 05:38:01.579247] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:28.247 [2024-12-07 05:38:01.579305] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:28.247 [2024-12-07 05:38:01.579324] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:28.247 [2024-12-07 05:38:01.581399] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:28.247 [2024-12-07 05:38:01.581453] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.247 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.529 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:28.529 "name": "Existed_Raid", 00:10:28.529 "uuid": "5c28520b-7b91-49fa-a303-ba1a7e400889", 00:10:28.529 "strip_size_kb": 0, 00:10:28.529 "state": "configuring", 00:10:28.529 "raid_level": "raid1", 00:10:28.529 "superblock": true, 00:10:28.529 "num_base_bdevs": 4, 00:10:28.529 "num_base_bdevs_discovered": 3, 00:10:28.529 "num_base_bdevs_operational": 4, 00:10:28.529 "base_bdevs_list": [ 00:10:28.529 { 00:10:28.529 "name": "BaseBdev1", 00:10:28.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:28.529 "is_configured": false, 00:10:28.529 "data_offset": 0, 00:10:28.529 "data_size": 0 00:10:28.529 }, 00:10:28.529 { 00:10:28.529 "name": "BaseBdev2", 00:10:28.529 "uuid": "1a75b823-bf80-4661-8b93-e11702fdd45c", 00:10:28.529 "is_configured": true, 00:10:28.529 "data_offset": 2048, 00:10:28.529 "data_size": 63488 00:10:28.529 }, 00:10:28.529 { 00:10:28.529 "name": "BaseBdev3", 00:10:28.529 "uuid": "d7f13a84-a8af-490d-b5cb-89370d72595a", 00:10:28.529 "is_configured": true, 00:10:28.529 "data_offset": 2048, 00:10:28.529 "data_size": 63488 00:10:28.529 }, 00:10:28.529 { 00:10:28.529 "name": "BaseBdev4", 00:10:28.529 "uuid": "2d15a1b8-b660-422d-890f-ff60dbaf8961", 00:10:28.529 "is_configured": true, 00:10:28.529 "data_offset": 2048, 00:10:28.529 "data_size": 63488 00:10:28.529 } 00:10:28.529 ] 00:10:28.529 }' 00:10:28.529 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:28.529 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.788 [2024-12-07 05:38:01.990640] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:28.788 05:38:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.788 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.788 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.788 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:28.788 "name": "Existed_Raid", 00:10:28.788 "uuid": "5c28520b-7b91-49fa-a303-ba1a7e400889", 00:10:28.788 "strip_size_kb": 0, 00:10:28.788 "state": "configuring", 00:10:28.788 "raid_level": "raid1", 00:10:28.788 "superblock": true, 00:10:28.788 "num_base_bdevs": 4, 00:10:28.788 "num_base_bdevs_discovered": 2, 00:10:28.788 "num_base_bdevs_operational": 4, 00:10:28.788 "base_bdevs_list": [ 00:10:28.788 { 00:10:28.788 "name": "BaseBdev1", 00:10:28.788 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:28.788 "is_configured": false, 00:10:28.788 "data_offset": 0, 00:10:28.788 "data_size": 0 00:10:28.788 }, 00:10:28.788 { 00:10:28.788 "name": null, 00:10:28.788 "uuid": "1a75b823-bf80-4661-8b93-e11702fdd45c", 00:10:28.788 "is_configured": false, 00:10:28.788 "data_offset": 0, 00:10:28.788 "data_size": 63488 00:10:28.788 }, 00:10:28.788 { 00:10:28.788 "name": "BaseBdev3", 00:10:28.788 "uuid": "d7f13a84-a8af-490d-b5cb-89370d72595a", 00:10:28.788 "is_configured": true, 00:10:28.788 "data_offset": 2048, 00:10:28.788 "data_size": 63488 00:10:28.788 }, 00:10:28.788 { 00:10:28.788 "name": "BaseBdev4", 00:10:28.788 "uuid": "2d15a1b8-b660-422d-890f-ff60dbaf8961", 00:10:28.788 "is_configured": true, 00:10:28.788 "data_offset": 2048, 00:10:28.788 "data_size": 63488 00:10:28.788 } 00:10:28.788 ] 00:10:28.788 }' 00:10:28.788 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:28.788 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.356 [2024-12-07 05:38:02.506669] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:29.356 BaseBdev1 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.356 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.356 [ 00:10:29.356 { 00:10:29.357 "name": "BaseBdev1", 00:10:29.357 "aliases": [ 00:10:29.357 "885d157d-f665-4858-9f50-c86c49f97064" 00:10:29.357 ], 00:10:29.357 "product_name": "Malloc disk", 00:10:29.357 "block_size": 512, 00:10:29.357 "num_blocks": 65536, 00:10:29.357 "uuid": "885d157d-f665-4858-9f50-c86c49f97064", 00:10:29.357 "assigned_rate_limits": { 00:10:29.357 "rw_ios_per_sec": 0, 00:10:29.357 "rw_mbytes_per_sec": 0, 00:10:29.357 "r_mbytes_per_sec": 0, 00:10:29.357 "w_mbytes_per_sec": 0 00:10:29.357 }, 00:10:29.357 "claimed": true, 00:10:29.357 "claim_type": "exclusive_write", 00:10:29.357 "zoned": false, 00:10:29.357 "supported_io_types": { 00:10:29.357 "read": true, 00:10:29.357 "write": true, 00:10:29.357 "unmap": true, 00:10:29.357 "flush": true, 00:10:29.357 "reset": true, 00:10:29.357 "nvme_admin": false, 00:10:29.357 "nvme_io": false, 00:10:29.357 "nvme_io_md": false, 00:10:29.357 "write_zeroes": true, 00:10:29.357 "zcopy": true, 00:10:29.357 "get_zone_info": false, 00:10:29.357 "zone_management": false, 00:10:29.357 "zone_append": false, 00:10:29.357 "compare": false, 00:10:29.357 "compare_and_write": false, 00:10:29.357 "abort": true, 00:10:29.357 "seek_hole": false, 00:10:29.357 "seek_data": false, 00:10:29.357 "copy": true, 00:10:29.357 "nvme_iov_md": false 00:10:29.357 }, 00:10:29.357 "memory_domains": [ 00:10:29.357 { 00:10:29.357 "dma_device_id": "system", 00:10:29.357 "dma_device_type": 1 00:10:29.357 }, 00:10:29.357 { 00:10:29.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:29.357 "dma_device_type": 2 00:10:29.357 } 00:10:29.357 ], 00:10:29.357 "driver_specific": {} 00:10:29.357 } 00:10:29.357 ] 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.357 "name": "Existed_Raid", 00:10:29.357 "uuid": "5c28520b-7b91-49fa-a303-ba1a7e400889", 00:10:29.357 "strip_size_kb": 0, 00:10:29.357 "state": "configuring", 00:10:29.357 "raid_level": "raid1", 00:10:29.357 "superblock": true, 00:10:29.357 "num_base_bdevs": 4, 00:10:29.357 "num_base_bdevs_discovered": 3, 00:10:29.357 "num_base_bdevs_operational": 4, 00:10:29.357 "base_bdevs_list": [ 00:10:29.357 { 00:10:29.357 "name": "BaseBdev1", 00:10:29.357 "uuid": "885d157d-f665-4858-9f50-c86c49f97064", 00:10:29.357 "is_configured": true, 00:10:29.357 "data_offset": 2048, 00:10:29.357 "data_size": 63488 00:10:29.357 }, 00:10:29.357 { 00:10:29.357 "name": null, 00:10:29.357 "uuid": "1a75b823-bf80-4661-8b93-e11702fdd45c", 00:10:29.357 "is_configured": false, 00:10:29.357 "data_offset": 0, 00:10:29.357 "data_size": 63488 00:10:29.357 }, 00:10:29.357 { 00:10:29.357 "name": "BaseBdev3", 00:10:29.357 "uuid": "d7f13a84-a8af-490d-b5cb-89370d72595a", 00:10:29.357 "is_configured": true, 00:10:29.357 "data_offset": 2048, 00:10:29.357 "data_size": 63488 00:10:29.357 }, 00:10:29.357 { 00:10:29.357 "name": "BaseBdev4", 00:10:29.357 "uuid": "2d15a1b8-b660-422d-890f-ff60dbaf8961", 00:10:29.357 "is_configured": true, 00:10:29.357 "data_offset": 2048, 00:10:29.357 "data_size": 63488 00:10:29.357 } 00:10:29.357 ] 00:10:29.357 }' 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.357 05:38:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.925 [2024-12-07 05:38:03.045869] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:29.925 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.926 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.926 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.926 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.926 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.926 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.926 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.926 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:29.926 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.926 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.926 "name": "Existed_Raid", 00:10:29.926 "uuid": "5c28520b-7b91-49fa-a303-ba1a7e400889", 00:10:29.926 "strip_size_kb": 0, 00:10:29.926 "state": "configuring", 00:10:29.926 "raid_level": "raid1", 00:10:29.926 "superblock": true, 00:10:29.926 "num_base_bdevs": 4, 00:10:29.926 "num_base_bdevs_discovered": 2, 00:10:29.926 "num_base_bdevs_operational": 4, 00:10:29.926 "base_bdevs_list": [ 00:10:29.926 { 00:10:29.926 "name": "BaseBdev1", 00:10:29.926 "uuid": "885d157d-f665-4858-9f50-c86c49f97064", 00:10:29.926 "is_configured": true, 00:10:29.926 "data_offset": 2048, 00:10:29.926 "data_size": 63488 00:10:29.926 }, 00:10:29.926 { 00:10:29.926 "name": null, 00:10:29.926 "uuid": "1a75b823-bf80-4661-8b93-e11702fdd45c", 00:10:29.926 "is_configured": false, 00:10:29.926 "data_offset": 0, 00:10:29.926 "data_size": 63488 00:10:29.926 }, 00:10:29.926 { 00:10:29.926 "name": null, 00:10:29.926 "uuid": "d7f13a84-a8af-490d-b5cb-89370d72595a", 00:10:29.926 "is_configured": false, 00:10:29.926 "data_offset": 0, 00:10:29.926 "data_size": 63488 00:10:29.926 }, 00:10:29.926 { 00:10:29.926 "name": "BaseBdev4", 00:10:29.926 "uuid": "2d15a1b8-b660-422d-890f-ff60dbaf8961", 00:10:29.926 "is_configured": true, 00:10:29.926 "data_offset": 2048, 00:10:29.926 "data_size": 63488 00:10:29.926 } 00:10:29.926 ] 00:10:29.926 }' 00:10:29.926 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.926 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.186 [2024-12-07 05:38:03.537191] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:30.186 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.444 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.444 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:30.444 "name": "Existed_Raid", 00:10:30.444 "uuid": "5c28520b-7b91-49fa-a303-ba1a7e400889", 00:10:30.444 "strip_size_kb": 0, 00:10:30.444 "state": "configuring", 00:10:30.444 "raid_level": "raid1", 00:10:30.444 "superblock": true, 00:10:30.444 "num_base_bdevs": 4, 00:10:30.444 "num_base_bdevs_discovered": 3, 00:10:30.444 "num_base_bdevs_operational": 4, 00:10:30.444 "base_bdevs_list": [ 00:10:30.444 { 00:10:30.444 "name": "BaseBdev1", 00:10:30.444 "uuid": "885d157d-f665-4858-9f50-c86c49f97064", 00:10:30.444 "is_configured": true, 00:10:30.444 "data_offset": 2048, 00:10:30.444 "data_size": 63488 00:10:30.444 }, 00:10:30.444 { 00:10:30.444 "name": null, 00:10:30.444 "uuid": "1a75b823-bf80-4661-8b93-e11702fdd45c", 00:10:30.444 "is_configured": false, 00:10:30.444 "data_offset": 0, 00:10:30.444 "data_size": 63488 00:10:30.444 }, 00:10:30.444 { 00:10:30.444 "name": "BaseBdev3", 00:10:30.444 "uuid": "d7f13a84-a8af-490d-b5cb-89370d72595a", 00:10:30.444 "is_configured": true, 00:10:30.444 "data_offset": 2048, 00:10:30.444 "data_size": 63488 00:10:30.444 }, 00:10:30.444 { 00:10:30.444 "name": "BaseBdev4", 00:10:30.444 "uuid": "2d15a1b8-b660-422d-890f-ff60dbaf8961", 00:10:30.444 "is_configured": true, 00:10:30.444 "data_offset": 2048, 00:10:30.444 "data_size": 63488 00:10:30.444 } 00:10:30.444 ] 00:10:30.444 }' 00:10:30.444 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:30.444 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.702 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:30.702 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.702 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.702 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.702 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.702 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:30.702 05:38:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:30.702 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.702 05:38:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.702 [2024-12-07 05:38:03.992559] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.702 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:30.702 "name": "Existed_Raid", 00:10:30.702 "uuid": "5c28520b-7b91-49fa-a303-ba1a7e400889", 00:10:30.702 "strip_size_kb": 0, 00:10:30.702 "state": "configuring", 00:10:30.703 "raid_level": "raid1", 00:10:30.703 "superblock": true, 00:10:30.703 "num_base_bdevs": 4, 00:10:30.703 "num_base_bdevs_discovered": 2, 00:10:30.703 "num_base_bdevs_operational": 4, 00:10:30.703 "base_bdevs_list": [ 00:10:30.703 { 00:10:30.703 "name": null, 00:10:30.703 "uuid": "885d157d-f665-4858-9f50-c86c49f97064", 00:10:30.703 "is_configured": false, 00:10:30.703 "data_offset": 0, 00:10:30.703 "data_size": 63488 00:10:30.703 }, 00:10:30.703 { 00:10:30.703 "name": null, 00:10:30.703 "uuid": "1a75b823-bf80-4661-8b93-e11702fdd45c", 00:10:30.703 "is_configured": false, 00:10:30.703 "data_offset": 0, 00:10:30.703 "data_size": 63488 00:10:30.703 }, 00:10:30.703 { 00:10:30.703 "name": "BaseBdev3", 00:10:30.703 "uuid": "d7f13a84-a8af-490d-b5cb-89370d72595a", 00:10:30.703 "is_configured": true, 00:10:30.703 "data_offset": 2048, 00:10:30.703 "data_size": 63488 00:10:30.703 }, 00:10:30.703 { 00:10:30.703 "name": "BaseBdev4", 00:10:30.703 "uuid": "2d15a1b8-b660-422d-890f-ff60dbaf8961", 00:10:30.703 "is_configured": true, 00:10:30.703 "data_offset": 2048, 00:10:30.703 "data_size": 63488 00:10:30.703 } 00:10:30.703 ] 00:10:30.703 }' 00:10:30.703 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:30.703 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.272 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.273 [2024-12-07 05:38:04.487611] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:31.273 "name": "Existed_Raid", 00:10:31.273 "uuid": "5c28520b-7b91-49fa-a303-ba1a7e400889", 00:10:31.273 "strip_size_kb": 0, 00:10:31.273 "state": "configuring", 00:10:31.273 "raid_level": "raid1", 00:10:31.273 "superblock": true, 00:10:31.273 "num_base_bdevs": 4, 00:10:31.273 "num_base_bdevs_discovered": 3, 00:10:31.273 "num_base_bdevs_operational": 4, 00:10:31.273 "base_bdevs_list": [ 00:10:31.273 { 00:10:31.273 "name": null, 00:10:31.273 "uuid": "885d157d-f665-4858-9f50-c86c49f97064", 00:10:31.273 "is_configured": false, 00:10:31.273 "data_offset": 0, 00:10:31.273 "data_size": 63488 00:10:31.273 }, 00:10:31.273 { 00:10:31.273 "name": "BaseBdev2", 00:10:31.273 "uuid": "1a75b823-bf80-4661-8b93-e11702fdd45c", 00:10:31.273 "is_configured": true, 00:10:31.273 "data_offset": 2048, 00:10:31.273 "data_size": 63488 00:10:31.273 }, 00:10:31.273 { 00:10:31.273 "name": "BaseBdev3", 00:10:31.273 "uuid": "d7f13a84-a8af-490d-b5cb-89370d72595a", 00:10:31.273 "is_configured": true, 00:10:31.273 "data_offset": 2048, 00:10:31.273 "data_size": 63488 00:10:31.273 }, 00:10:31.273 { 00:10:31.273 "name": "BaseBdev4", 00:10:31.273 "uuid": "2d15a1b8-b660-422d-890f-ff60dbaf8961", 00:10:31.273 "is_configured": true, 00:10:31.273 "data_offset": 2048, 00:10:31.273 "data_size": 63488 00:10:31.273 } 00:10:31.273 ] 00:10:31.273 }' 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:31.273 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 885d157d-f665-4858-9f50-c86c49f97064 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.843 05:38:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.843 [2024-12-07 05:38:05.019708] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:31.843 [2024-12-07 05:38:05.019914] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:31.843 [2024-12-07 05:38:05.019932] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:31.843 [2024-12-07 05:38:05.020220] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:31.844 [2024-12-07 05:38:05.020381] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:31.844 [2024-12-07 05:38:05.020397] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:31.844 [2024-12-07 05:38:05.020515] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:31.844 NewBaseBdev 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.844 [ 00:10:31.844 { 00:10:31.844 "name": "NewBaseBdev", 00:10:31.844 "aliases": [ 00:10:31.844 "885d157d-f665-4858-9f50-c86c49f97064" 00:10:31.844 ], 00:10:31.844 "product_name": "Malloc disk", 00:10:31.844 "block_size": 512, 00:10:31.844 "num_blocks": 65536, 00:10:31.844 "uuid": "885d157d-f665-4858-9f50-c86c49f97064", 00:10:31.844 "assigned_rate_limits": { 00:10:31.844 "rw_ios_per_sec": 0, 00:10:31.844 "rw_mbytes_per_sec": 0, 00:10:31.844 "r_mbytes_per_sec": 0, 00:10:31.844 "w_mbytes_per_sec": 0 00:10:31.844 }, 00:10:31.844 "claimed": true, 00:10:31.844 "claim_type": "exclusive_write", 00:10:31.844 "zoned": false, 00:10:31.844 "supported_io_types": { 00:10:31.844 "read": true, 00:10:31.844 "write": true, 00:10:31.844 "unmap": true, 00:10:31.844 "flush": true, 00:10:31.844 "reset": true, 00:10:31.844 "nvme_admin": false, 00:10:31.844 "nvme_io": false, 00:10:31.844 "nvme_io_md": false, 00:10:31.844 "write_zeroes": true, 00:10:31.844 "zcopy": true, 00:10:31.844 "get_zone_info": false, 00:10:31.844 "zone_management": false, 00:10:31.844 "zone_append": false, 00:10:31.844 "compare": false, 00:10:31.844 "compare_and_write": false, 00:10:31.844 "abort": true, 00:10:31.844 "seek_hole": false, 00:10:31.844 "seek_data": false, 00:10:31.844 "copy": true, 00:10:31.844 "nvme_iov_md": false 00:10:31.844 }, 00:10:31.844 "memory_domains": [ 00:10:31.844 { 00:10:31.844 "dma_device_id": "system", 00:10:31.844 "dma_device_type": 1 00:10:31.844 }, 00:10:31.844 { 00:10:31.844 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:31.844 "dma_device_type": 2 00:10:31.844 } 00:10:31.844 ], 00:10:31.844 "driver_specific": {} 00:10:31.844 } 00:10:31.844 ] 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:31.844 "name": "Existed_Raid", 00:10:31.844 "uuid": "5c28520b-7b91-49fa-a303-ba1a7e400889", 00:10:31.844 "strip_size_kb": 0, 00:10:31.844 "state": "online", 00:10:31.844 "raid_level": "raid1", 00:10:31.844 "superblock": true, 00:10:31.844 "num_base_bdevs": 4, 00:10:31.844 "num_base_bdevs_discovered": 4, 00:10:31.844 "num_base_bdevs_operational": 4, 00:10:31.844 "base_bdevs_list": [ 00:10:31.844 { 00:10:31.844 "name": "NewBaseBdev", 00:10:31.844 "uuid": "885d157d-f665-4858-9f50-c86c49f97064", 00:10:31.844 "is_configured": true, 00:10:31.844 "data_offset": 2048, 00:10:31.844 "data_size": 63488 00:10:31.844 }, 00:10:31.844 { 00:10:31.844 "name": "BaseBdev2", 00:10:31.844 "uuid": "1a75b823-bf80-4661-8b93-e11702fdd45c", 00:10:31.844 "is_configured": true, 00:10:31.844 "data_offset": 2048, 00:10:31.844 "data_size": 63488 00:10:31.844 }, 00:10:31.844 { 00:10:31.844 "name": "BaseBdev3", 00:10:31.844 "uuid": "d7f13a84-a8af-490d-b5cb-89370d72595a", 00:10:31.844 "is_configured": true, 00:10:31.844 "data_offset": 2048, 00:10:31.844 "data_size": 63488 00:10:31.844 }, 00:10:31.844 { 00:10:31.844 "name": "BaseBdev4", 00:10:31.844 "uuid": "2d15a1b8-b660-422d-890f-ff60dbaf8961", 00:10:31.844 "is_configured": true, 00:10:31.844 "data_offset": 2048, 00:10:31.844 "data_size": 63488 00:10:31.844 } 00:10:31.844 ] 00:10:31.844 }' 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:31.844 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.413 [2024-12-07 05:38:05.547211] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:32.413 "name": "Existed_Raid", 00:10:32.413 "aliases": [ 00:10:32.413 "5c28520b-7b91-49fa-a303-ba1a7e400889" 00:10:32.413 ], 00:10:32.413 "product_name": "Raid Volume", 00:10:32.413 "block_size": 512, 00:10:32.413 "num_blocks": 63488, 00:10:32.413 "uuid": "5c28520b-7b91-49fa-a303-ba1a7e400889", 00:10:32.413 "assigned_rate_limits": { 00:10:32.413 "rw_ios_per_sec": 0, 00:10:32.413 "rw_mbytes_per_sec": 0, 00:10:32.413 "r_mbytes_per_sec": 0, 00:10:32.413 "w_mbytes_per_sec": 0 00:10:32.413 }, 00:10:32.413 "claimed": false, 00:10:32.413 "zoned": false, 00:10:32.413 "supported_io_types": { 00:10:32.413 "read": true, 00:10:32.413 "write": true, 00:10:32.413 "unmap": false, 00:10:32.413 "flush": false, 00:10:32.413 "reset": true, 00:10:32.413 "nvme_admin": false, 00:10:32.413 "nvme_io": false, 00:10:32.413 "nvme_io_md": false, 00:10:32.413 "write_zeroes": true, 00:10:32.413 "zcopy": false, 00:10:32.413 "get_zone_info": false, 00:10:32.413 "zone_management": false, 00:10:32.413 "zone_append": false, 00:10:32.413 "compare": false, 00:10:32.413 "compare_and_write": false, 00:10:32.413 "abort": false, 00:10:32.413 "seek_hole": false, 00:10:32.413 "seek_data": false, 00:10:32.413 "copy": false, 00:10:32.413 "nvme_iov_md": false 00:10:32.413 }, 00:10:32.413 "memory_domains": [ 00:10:32.413 { 00:10:32.413 "dma_device_id": "system", 00:10:32.413 "dma_device_type": 1 00:10:32.413 }, 00:10:32.413 { 00:10:32.413 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.413 "dma_device_type": 2 00:10:32.413 }, 00:10:32.413 { 00:10:32.413 "dma_device_id": "system", 00:10:32.413 "dma_device_type": 1 00:10:32.413 }, 00:10:32.413 { 00:10:32.413 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.413 "dma_device_type": 2 00:10:32.413 }, 00:10:32.413 { 00:10:32.413 "dma_device_id": "system", 00:10:32.413 "dma_device_type": 1 00:10:32.413 }, 00:10:32.413 { 00:10:32.413 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.413 "dma_device_type": 2 00:10:32.413 }, 00:10:32.413 { 00:10:32.413 "dma_device_id": "system", 00:10:32.413 "dma_device_type": 1 00:10:32.413 }, 00:10:32.413 { 00:10:32.413 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.413 "dma_device_type": 2 00:10:32.413 } 00:10:32.413 ], 00:10:32.413 "driver_specific": { 00:10:32.413 "raid": { 00:10:32.413 "uuid": "5c28520b-7b91-49fa-a303-ba1a7e400889", 00:10:32.413 "strip_size_kb": 0, 00:10:32.413 "state": "online", 00:10:32.413 "raid_level": "raid1", 00:10:32.413 "superblock": true, 00:10:32.413 "num_base_bdevs": 4, 00:10:32.413 "num_base_bdevs_discovered": 4, 00:10:32.413 "num_base_bdevs_operational": 4, 00:10:32.413 "base_bdevs_list": [ 00:10:32.413 { 00:10:32.413 "name": "NewBaseBdev", 00:10:32.413 "uuid": "885d157d-f665-4858-9f50-c86c49f97064", 00:10:32.413 "is_configured": true, 00:10:32.413 "data_offset": 2048, 00:10:32.413 "data_size": 63488 00:10:32.413 }, 00:10:32.413 { 00:10:32.413 "name": "BaseBdev2", 00:10:32.413 "uuid": "1a75b823-bf80-4661-8b93-e11702fdd45c", 00:10:32.413 "is_configured": true, 00:10:32.413 "data_offset": 2048, 00:10:32.413 "data_size": 63488 00:10:32.413 }, 00:10:32.413 { 00:10:32.413 "name": "BaseBdev3", 00:10:32.413 "uuid": "d7f13a84-a8af-490d-b5cb-89370d72595a", 00:10:32.413 "is_configured": true, 00:10:32.413 "data_offset": 2048, 00:10:32.413 "data_size": 63488 00:10:32.413 }, 00:10:32.413 { 00:10:32.413 "name": "BaseBdev4", 00:10:32.413 "uuid": "2d15a1b8-b660-422d-890f-ff60dbaf8961", 00:10:32.413 "is_configured": true, 00:10:32.413 "data_offset": 2048, 00:10:32.413 "data_size": 63488 00:10:32.413 } 00:10:32.413 ] 00:10:32.413 } 00:10:32.413 } 00:10:32.413 }' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:32.413 BaseBdev2 00:10:32.413 BaseBdev3 00:10:32.413 BaseBdev4' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.413 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.414 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.414 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.414 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:32.414 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.414 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.414 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.414 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.414 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.414 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:32.414 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.414 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.674 [2024-12-07 05:38:05.782460] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:32.674 [2024-12-07 05:38:05.782513] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:32.674 [2024-12-07 05:38:05.782608] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:32.674 [2024-12-07 05:38:05.782911] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:32.674 [2024-12-07 05:38:05.782936] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 84278 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 84278 ']' 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 84278 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84278 00:10:32.674 killing process with pid 84278 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84278' 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 84278 00:10:32.674 [2024-12-07 05:38:05.826283] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:32.674 05:38:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 84278 00:10:32.674 [2024-12-07 05:38:05.905178] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:32.933 05:38:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:32.933 00:10:32.933 real 0m9.558s 00:10:32.933 user 0m16.063s 00:10:32.933 sys 0m2.048s 00:10:32.933 05:38:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:32.933 05:38:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.933 ************************************ 00:10:32.933 END TEST raid_state_function_test_sb 00:10:32.933 ************************************ 00:10:32.933 05:38:06 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:10:32.933 05:38:06 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:10:32.933 05:38:06 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:32.933 05:38:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:33.194 ************************************ 00:10:33.194 START TEST raid_superblock_test 00:10:33.194 ************************************ 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 4 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=84932 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 84932 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 84932 ']' 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:33.194 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:33.194 05:38:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.194 [2024-12-07 05:38:06.387203] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:10:33.194 [2024-12-07 05:38:06.387343] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84932 ] 00:10:33.194 [2024-12-07 05:38:06.543754] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:33.454 [2024-12-07 05:38:06.583530] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:33.454 [2024-12-07 05:38:06.659981] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:33.454 [2024-12-07 05:38:06.660025] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.024 malloc1 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.024 [2024-12-07 05:38:07.253705] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:34.024 [2024-12-07 05:38:07.253780] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:34.024 [2024-12-07 05:38:07.253804] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:34.024 [2024-12-07 05:38:07.253821] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:34.024 [2024-12-07 05:38:07.256263] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:34.024 [2024-12-07 05:38:07.256302] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:34.024 pt1 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.024 malloc2 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.024 [2024-12-07 05:38:07.288451] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:34.024 [2024-12-07 05:38:07.288509] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:34.024 [2024-12-07 05:38:07.288529] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:34.024 [2024-12-07 05:38:07.288541] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:34.024 [2024-12-07 05:38:07.290989] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:34.024 [2024-12-07 05:38:07.291022] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:34.024 pt2 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.024 malloc3 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.024 [2024-12-07 05:38:07.323147] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:34.024 [2024-12-07 05:38:07.323201] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:34.024 [2024-12-07 05:38:07.323220] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:34.024 [2024-12-07 05:38:07.323232] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:34.024 [2024-12-07 05:38:07.325645] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:34.024 [2024-12-07 05:38:07.325678] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:34.024 pt3 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.024 malloc4 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.024 [2024-12-07 05:38:07.367849] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:34.024 [2024-12-07 05:38:07.367899] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:34.024 [2024-12-07 05:38:07.367915] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:34.024 [2024-12-07 05:38:07.367930] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:34.024 [2024-12-07 05:38:07.370282] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:34.024 [2024-12-07 05:38:07.370316] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:34.024 pt4 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.024 [2024-12-07 05:38:07.379858] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:34.024 [2024-12-07 05:38:07.382006] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:34.024 [2024-12-07 05:38:07.382079] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:34.024 [2024-12-07 05:38:07.382147] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:34.024 [2024-12-07 05:38:07.382314] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:34.024 [2024-12-07 05:38:07.382335] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:34.024 [2024-12-07 05:38:07.382615] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:34.024 [2024-12-07 05:38:07.382815] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:34.024 [2024-12-07 05:38:07.382831] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:34.024 [2024-12-07 05:38:07.382951] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.024 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.285 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.285 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.285 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:34.285 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.285 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.285 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.285 "name": "raid_bdev1", 00:10:34.285 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:34.285 "strip_size_kb": 0, 00:10:34.285 "state": "online", 00:10:34.285 "raid_level": "raid1", 00:10:34.285 "superblock": true, 00:10:34.285 "num_base_bdevs": 4, 00:10:34.285 "num_base_bdevs_discovered": 4, 00:10:34.285 "num_base_bdevs_operational": 4, 00:10:34.285 "base_bdevs_list": [ 00:10:34.285 { 00:10:34.285 "name": "pt1", 00:10:34.285 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:34.285 "is_configured": true, 00:10:34.285 "data_offset": 2048, 00:10:34.285 "data_size": 63488 00:10:34.285 }, 00:10:34.285 { 00:10:34.285 "name": "pt2", 00:10:34.285 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:34.285 "is_configured": true, 00:10:34.285 "data_offset": 2048, 00:10:34.285 "data_size": 63488 00:10:34.285 }, 00:10:34.285 { 00:10:34.285 "name": "pt3", 00:10:34.285 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:34.285 "is_configured": true, 00:10:34.285 "data_offset": 2048, 00:10:34.285 "data_size": 63488 00:10:34.285 }, 00:10:34.285 { 00:10:34.285 "name": "pt4", 00:10:34.285 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:34.285 "is_configured": true, 00:10:34.285 "data_offset": 2048, 00:10:34.285 "data_size": 63488 00:10:34.285 } 00:10:34.285 ] 00:10:34.285 }' 00:10:34.285 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.285 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:34.545 [2024-12-07 05:38:07.803552] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:34.545 "name": "raid_bdev1", 00:10:34.545 "aliases": [ 00:10:34.545 "529b36b2-ea95-452d-8bf6-d82bb7f31c7e" 00:10:34.545 ], 00:10:34.545 "product_name": "Raid Volume", 00:10:34.545 "block_size": 512, 00:10:34.545 "num_blocks": 63488, 00:10:34.545 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:34.545 "assigned_rate_limits": { 00:10:34.545 "rw_ios_per_sec": 0, 00:10:34.545 "rw_mbytes_per_sec": 0, 00:10:34.545 "r_mbytes_per_sec": 0, 00:10:34.545 "w_mbytes_per_sec": 0 00:10:34.545 }, 00:10:34.545 "claimed": false, 00:10:34.545 "zoned": false, 00:10:34.545 "supported_io_types": { 00:10:34.545 "read": true, 00:10:34.545 "write": true, 00:10:34.545 "unmap": false, 00:10:34.545 "flush": false, 00:10:34.545 "reset": true, 00:10:34.545 "nvme_admin": false, 00:10:34.545 "nvme_io": false, 00:10:34.545 "nvme_io_md": false, 00:10:34.545 "write_zeroes": true, 00:10:34.545 "zcopy": false, 00:10:34.545 "get_zone_info": false, 00:10:34.545 "zone_management": false, 00:10:34.545 "zone_append": false, 00:10:34.545 "compare": false, 00:10:34.545 "compare_and_write": false, 00:10:34.545 "abort": false, 00:10:34.545 "seek_hole": false, 00:10:34.545 "seek_data": false, 00:10:34.545 "copy": false, 00:10:34.545 "nvme_iov_md": false 00:10:34.545 }, 00:10:34.545 "memory_domains": [ 00:10:34.545 { 00:10:34.545 "dma_device_id": "system", 00:10:34.545 "dma_device_type": 1 00:10:34.545 }, 00:10:34.545 { 00:10:34.545 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.545 "dma_device_type": 2 00:10:34.545 }, 00:10:34.545 { 00:10:34.545 "dma_device_id": "system", 00:10:34.545 "dma_device_type": 1 00:10:34.545 }, 00:10:34.545 { 00:10:34.545 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.545 "dma_device_type": 2 00:10:34.545 }, 00:10:34.545 { 00:10:34.545 "dma_device_id": "system", 00:10:34.545 "dma_device_type": 1 00:10:34.545 }, 00:10:34.545 { 00:10:34.545 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.545 "dma_device_type": 2 00:10:34.545 }, 00:10:34.545 { 00:10:34.545 "dma_device_id": "system", 00:10:34.545 "dma_device_type": 1 00:10:34.545 }, 00:10:34.545 { 00:10:34.545 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.545 "dma_device_type": 2 00:10:34.545 } 00:10:34.545 ], 00:10:34.545 "driver_specific": { 00:10:34.545 "raid": { 00:10:34.545 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:34.545 "strip_size_kb": 0, 00:10:34.545 "state": "online", 00:10:34.545 "raid_level": "raid1", 00:10:34.545 "superblock": true, 00:10:34.545 "num_base_bdevs": 4, 00:10:34.545 "num_base_bdevs_discovered": 4, 00:10:34.545 "num_base_bdevs_operational": 4, 00:10:34.545 "base_bdevs_list": [ 00:10:34.545 { 00:10:34.545 "name": "pt1", 00:10:34.545 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:34.545 "is_configured": true, 00:10:34.545 "data_offset": 2048, 00:10:34.545 "data_size": 63488 00:10:34.545 }, 00:10:34.545 { 00:10:34.545 "name": "pt2", 00:10:34.545 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:34.545 "is_configured": true, 00:10:34.545 "data_offset": 2048, 00:10:34.545 "data_size": 63488 00:10:34.545 }, 00:10:34.545 { 00:10:34.545 "name": "pt3", 00:10:34.545 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:34.545 "is_configured": true, 00:10:34.545 "data_offset": 2048, 00:10:34.545 "data_size": 63488 00:10:34.545 }, 00:10:34.545 { 00:10:34.545 "name": "pt4", 00:10:34.545 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:34.545 "is_configured": true, 00:10:34.545 "data_offset": 2048, 00:10:34.545 "data_size": 63488 00:10:34.545 } 00:10:34.545 ] 00:10:34.545 } 00:10:34.545 } 00:10:34.545 }' 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:34.545 pt2 00:10:34.545 pt3 00:10:34.545 pt4' 00:10:34.545 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.807 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:34.807 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.807 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:34.808 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.808 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.808 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.808 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.808 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:34.808 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:34.808 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.808 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.808 05:38:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:34.808 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.808 05:38:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.808 [2024-12-07 05:38:08.142896] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:34.808 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=529b36b2-ea95-452d-8bf6-d82bb7f31c7e 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 529b36b2-ea95-452d-8bf6-d82bb7f31c7e ']' 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.068 [2024-12-07 05:38:08.190577] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:35.068 [2024-12-07 05:38:08.190615] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:35.068 [2024-12-07 05:38:08.190720] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:35.068 [2024-12-07 05:38:08.190869] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:35.068 [2024-12-07 05:38:08.190881] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.068 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.069 [2024-12-07 05:38:08.358692] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:35.069 [2024-12-07 05:38:08.360862] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:35.069 [2024-12-07 05:38:08.360920] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:35.069 [2024-12-07 05:38:08.360951] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:35.069 [2024-12-07 05:38:08.361009] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:35.069 [2024-12-07 05:38:08.361062] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:35.069 [2024-12-07 05:38:08.361081] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:35.069 [2024-12-07 05:38:08.361098] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:35.069 [2024-12-07 05:38:08.361113] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:35.069 [2024-12-07 05:38:08.361124] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:10:35.069 request: 00:10:35.069 { 00:10:35.069 "name": "raid_bdev1", 00:10:35.069 "raid_level": "raid1", 00:10:35.069 "base_bdevs": [ 00:10:35.069 "malloc1", 00:10:35.069 "malloc2", 00:10:35.069 "malloc3", 00:10:35.069 "malloc4" 00:10:35.069 ], 00:10:35.069 "superblock": false, 00:10:35.069 "method": "bdev_raid_create", 00:10:35.069 "req_id": 1 00:10:35.069 } 00:10:35.069 Got JSON-RPC error response 00:10:35.069 response: 00:10:35.069 { 00:10:35.069 "code": -17, 00:10:35.069 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:35.069 } 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.069 [2024-12-07 05:38:08.422517] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:35.069 [2024-12-07 05:38:08.422608] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:35.069 [2024-12-07 05:38:08.422649] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:35.069 [2024-12-07 05:38:08.422661] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:35.069 [2024-12-07 05:38:08.425164] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:35.069 [2024-12-07 05:38:08.425198] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:35.069 [2024-12-07 05:38:08.425303] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:35.069 [2024-12-07 05:38:08.425346] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:35.069 pt1 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.069 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.329 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.330 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:35.330 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.330 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.330 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.330 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.330 "name": "raid_bdev1", 00:10:35.330 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:35.330 "strip_size_kb": 0, 00:10:35.330 "state": "configuring", 00:10:35.330 "raid_level": "raid1", 00:10:35.330 "superblock": true, 00:10:35.330 "num_base_bdevs": 4, 00:10:35.330 "num_base_bdevs_discovered": 1, 00:10:35.330 "num_base_bdevs_operational": 4, 00:10:35.330 "base_bdevs_list": [ 00:10:35.330 { 00:10:35.330 "name": "pt1", 00:10:35.330 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:35.330 "is_configured": true, 00:10:35.330 "data_offset": 2048, 00:10:35.330 "data_size": 63488 00:10:35.330 }, 00:10:35.330 { 00:10:35.330 "name": null, 00:10:35.330 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:35.330 "is_configured": false, 00:10:35.330 "data_offset": 2048, 00:10:35.330 "data_size": 63488 00:10:35.330 }, 00:10:35.330 { 00:10:35.330 "name": null, 00:10:35.330 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:35.330 "is_configured": false, 00:10:35.330 "data_offset": 2048, 00:10:35.330 "data_size": 63488 00:10:35.330 }, 00:10:35.330 { 00:10:35.330 "name": null, 00:10:35.330 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:35.330 "is_configured": false, 00:10:35.330 "data_offset": 2048, 00:10:35.330 "data_size": 63488 00:10:35.330 } 00:10:35.330 ] 00:10:35.330 }' 00:10:35.330 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.330 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.591 [2024-12-07 05:38:08.865710] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:35.591 [2024-12-07 05:38:08.865817] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:35.591 [2024-12-07 05:38:08.865843] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:10:35.591 [2024-12-07 05:38:08.865854] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:35.591 [2024-12-07 05:38:08.866364] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:35.591 [2024-12-07 05:38:08.866384] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:35.591 [2024-12-07 05:38:08.866523] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:35.591 [2024-12-07 05:38:08.866556] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:35.591 pt2 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.591 [2024-12-07 05:38:08.877807] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.591 "name": "raid_bdev1", 00:10:35.591 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:35.591 "strip_size_kb": 0, 00:10:35.591 "state": "configuring", 00:10:35.591 "raid_level": "raid1", 00:10:35.591 "superblock": true, 00:10:35.591 "num_base_bdevs": 4, 00:10:35.591 "num_base_bdevs_discovered": 1, 00:10:35.591 "num_base_bdevs_operational": 4, 00:10:35.591 "base_bdevs_list": [ 00:10:35.591 { 00:10:35.591 "name": "pt1", 00:10:35.591 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:35.591 "is_configured": true, 00:10:35.591 "data_offset": 2048, 00:10:35.591 "data_size": 63488 00:10:35.591 }, 00:10:35.591 { 00:10:35.591 "name": null, 00:10:35.591 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:35.591 "is_configured": false, 00:10:35.591 "data_offset": 0, 00:10:35.591 "data_size": 63488 00:10:35.591 }, 00:10:35.591 { 00:10:35.591 "name": null, 00:10:35.591 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:35.591 "is_configured": false, 00:10:35.591 "data_offset": 2048, 00:10:35.591 "data_size": 63488 00:10:35.591 }, 00:10:35.591 { 00:10:35.591 "name": null, 00:10:35.591 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:35.591 "is_configured": false, 00:10:35.591 "data_offset": 2048, 00:10:35.591 "data_size": 63488 00:10:35.591 } 00:10:35.591 ] 00:10:35.591 }' 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.591 05:38:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.163 [2024-12-07 05:38:09.312936] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:36.163 [2024-12-07 05:38:09.313031] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:36.163 [2024-12-07 05:38:09.313051] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:36.163 [2024-12-07 05:38:09.313064] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:36.163 [2024-12-07 05:38:09.313542] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:36.163 [2024-12-07 05:38:09.313571] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:36.163 [2024-12-07 05:38:09.313675] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:36.163 [2024-12-07 05:38:09.313710] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:36.163 pt2 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.163 [2024-12-07 05:38:09.320848] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:36.163 [2024-12-07 05:38:09.320899] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:36.163 [2024-12-07 05:38:09.320917] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:36.163 [2024-12-07 05:38:09.320929] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:36.163 [2024-12-07 05:38:09.321334] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:36.163 [2024-12-07 05:38:09.321354] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:36.163 [2024-12-07 05:38:09.321414] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:36.163 [2024-12-07 05:38:09.321446] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:36.163 pt3 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.163 [2024-12-07 05:38:09.328844] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:36.163 [2024-12-07 05:38:09.328892] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:36.163 [2024-12-07 05:38:09.328907] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:10:36.163 [2024-12-07 05:38:09.328917] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:36.163 [2024-12-07 05:38:09.329294] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:36.163 [2024-12-07 05:38:09.329324] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:36.163 [2024-12-07 05:38:09.329390] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:36.163 [2024-12-07 05:38:09.329414] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:36.163 [2024-12-07 05:38:09.329546] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:36.163 [2024-12-07 05:38:09.329571] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:36.163 [2024-12-07 05:38:09.329865] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:36.163 [2024-12-07 05:38:09.330022] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:36.163 [2024-12-07 05:38:09.330033] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:10:36.163 [2024-12-07 05:38:09.330153] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:36.163 pt4 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.163 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.163 "name": "raid_bdev1", 00:10:36.163 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:36.163 "strip_size_kb": 0, 00:10:36.163 "state": "online", 00:10:36.163 "raid_level": "raid1", 00:10:36.163 "superblock": true, 00:10:36.163 "num_base_bdevs": 4, 00:10:36.163 "num_base_bdevs_discovered": 4, 00:10:36.163 "num_base_bdevs_operational": 4, 00:10:36.163 "base_bdevs_list": [ 00:10:36.163 { 00:10:36.163 "name": "pt1", 00:10:36.163 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:36.163 "is_configured": true, 00:10:36.163 "data_offset": 2048, 00:10:36.163 "data_size": 63488 00:10:36.163 }, 00:10:36.163 { 00:10:36.163 "name": "pt2", 00:10:36.163 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:36.163 "is_configured": true, 00:10:36.163 "data_offset": 2048, 00:10:36.163 "data_size": 63488 00:10:36.163 }, 00:10:36.163 { 00:10:36.163 "name": "pt3", 00:10:36.163 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:36.163 "is_configured": true, 00:10:36.164 "data_offset": 2048, 00:10:36.164 "data_size": 63488 00:10:36.164 }, 00:10:36.164 { 00:10:36.164 "name": "pt4", 00:10:36.164 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:36.164 "is_configured": true, 00:10:36.164 "data_offset": 2048, 00:10:36.164 "data_size": 63488 00:10:36.164 } 00:10:36.164 ] 00:10:36.164 }' 00:10:36.164 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.164 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.736 [2024-12-07 05:38:09.812421] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:36.736 "name": "raid_bdev1", 00:10:36.736 "aliases": [ 00:10:36.736 "529b36b2-ea95-452d-8bf6-d82bb7f31c7e" 00:10:36.736 ], 00:10:36.736 "product_name": "Raid Volume", 00:10:36.736 "block_size": 512, 00:10:36.736 "num_blocks": 63488, 00:10:36.736 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:36.736 "assigned_rate_limits": { 00:10:36.736 "rw_ios_per_sec": 0, 00:10:36.736 "rw_mbytes_per_sec": 0, 00:10:36.736 "r_mbytes_per_sec": 0, 00:10:36.736 "w_mbytes_per_sec": 0 00:10:36.736 }, 00:10:36.736 "claimed": false, 00:10:36.736 "zoned": false, 00:10:36.736 "supported_io_types": { 00:10:36.736 "read": true, 00:10:36.736 "write": true, 00:10:36.736 "unmap": false, 00:10:36.736 "flush": false, 00:10:36.736 "reset": true, 00:10:36.736 "nvme_admin": false, 00:10:36.736 "nvme_io": false, 00:10:36.736 "nvme_io_md": false, 00:10:36.736 "write_zeroes": true, 00:10:36.736 "zcopy": false, 00:10:36.736 "get_zone_info": false, 00:10:36.736 "zone_management": false, 00:10:36.736 "zone_append": false, 00:10:36.736 "compare": false, 00:10:36.736 "compare_and_write": false, 00:10:36.736 "abort": false, 00:10:36.736 "seek_hole": false, 00:10:36.736 "seek_data": false, 00:10:36.736 "copy": false, 00:10:36.736 "nvme_iov_md": false 00:10:36.736 }, 00:10:36.736 "memory_domains": [ 00:10:36.736 { 00:10:36.736 "dma_device_id": "system", 00:10:36.736 "dma_device_type": 1 00:10:36.736 }, 00:10:36.736 { 00:10:36.736 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:36.736 "dma_device_type": 2 00:10:36.736 }, 00:10:36.736 { 00:10:36.736 "dma_device_id": "system", 00:10:36.736 "dma_device_type": 1 00:10:36.736 }, 00:10:36.736 { 00:10:36.736 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:36.736 "dma_device_type": 2 00:10:36.736 }, 00:10:36.736 { 00:10:36.736 "dma_device_id": "system", 00:10:36.736 "dma_device_type": 1 00:10:36.736 }, 00:10:36.736 { 00:10:36.736 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:36.736 "dma_device_type": 2 00:10:36.736 }, 00:10:36.736 { 00:10:36.736 "dma_device_id": "system", 00:10:36.736 "dma_device_type": 1 00:10:36.736 }, 00:10:36.736 { 00:10:36.736 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:36.736 "dma_device_type": 2 00:10:36.736 } 00:10:36.736 ], 00:10:36.736 "driver_specific": { 00:10:36.736 "raid": { 00:10:36.736 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:36.736 "strip_size_kb": 0, 00:10:36.736 "state": "online", 00:10:36.736 "raid_level": "raid1", 00:10:36.736 "superblock": true, 00:10:36.736 "num_base_bdevs": 4, 00:10:36.736 "num_base_bdevs_discovered": 4, 00:10:36.736 "num_base_bdevs_operational": 4, 00:10:36.736 "base_bdevs_list": [ 00:10:36.736 { 00:10:36.736 "name": "pt1", 00:10:36.736 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:36.736 "is_configured": true, 00:10:36.736 "data_offset": 2048, 00:10:36.736 "data_size": 63488 00:10:36.736 }, 00:10:36.736 { 00:10:36.736 "name": "pt2", 00:10:36.736 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:36.736 "is_configured": true, 00:10:36.736 "data_offset": 2048, 00:10:36.736 "data_size": 63488 00:10:36.736 }, 00:10:36.736 { 00:10:36.736 "name": "pt3", 00:10:36.736 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:36.736 "is_configured": true, 00:10:36.736 "data_offset": 2048, 00:10:36.736 "data_size": 63488 00:10:36.736 }, 00:10:36.736 { 00:10:36.736 "name": "pt4", 00:10:36.736 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:36.736 "is_configured": true, 00:10:36.736 "data_offset": 2048, 00:10:36.736 "data_size": 63488 00:10:36.736 } 00:10:36.736 ] 00:10:36.736 } 00:10:36.736 } 00:10:36.736 }' 00:10:36.736 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:36.737 pt2 00:10:36.737 pt3 00:10:36.737 pt4' 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.737 05:38:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.737 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.997 [2024-12-07 05:38:10.151801] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 529b36b2-ea95-452d-8bf6-d82bb7f31c7e '!=' 529b36b2-ea95-452d-8bf6-d82bb7f31c7e ']' 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.997 [2024-12-07 05:38:10.195457] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.997 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.997 "name": "raid_bdev1", 00:10:36.997 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:36.997 "strip_size_kb": 0, 00:10:36.997 "state": "online", 00:10:36.997 "raid_level": "raid1", 00:10:36.997 "superblock": true, 00:10:36.997 "num_base_bdevs": 4, 00:10:36.997 "num_base_bdevs_discovered": 3, 00:10:36.997 "num_base_bdevs_operational": 3, 00:10:36.997 "base_bdevs_list": [ 00:10:36.997 { 00:10:36.997 "name": null, 00:10:36.997 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.997 "is_configured": false, 00:10:36.997 "data_offset": 0, 00:10:36.998 "data_size": 63488 00:10:36.998 }, 00:10:36.998 { 00:10:36.998 "name": "pt2", 00:10:36.998 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:36.998 "is_configured": true, 00:10:36.998 "data_offset": 2048, 00:10:36.998 "data_size": 63488 00:10:36.998 }, 00:10:36.998 { 00:10:36.998 "name": "pt3", 00:10:36.998 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:36.998 "is_configured": true, 00:10:36.998 "data_offset": 2048, 00:10:36.998 "data_size": 63488 00:10:36.998 }, 00:10:36.998 { 00:10:36.998 "name": "pt4", 00:10:36.998 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:36.998 "is_configured": true, 00:10:36.998 "data_offset": 2048, 00:10:36.998 "data_size": 63488 00:10:36.998 } 00:10:36.998 ] 00:10:36.998 }' 00:10:36.998 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.998 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.258 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:37.258 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.258 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.258 [2024-12-07 05:38:10.610818] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:37.258 [2024-12-07 05:38:10.610869] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:37.258 [2024-12-07 05:38:10.610989] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:37.258 [2024-12-07 05:38:10.611083] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:37.258 [2024-12-07 05:38:10.611101] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:10:37.258 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.258 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.258 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.258 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.258 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.519 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.519 [2024-12-07 05:38:10.706586] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:37.519 [2024-12-07 05:38:10.706660] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:37.519 [2024-12-07 05:38:10.706678] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:37.519 [2024-12-07 05:38:10.706690] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:37.519 [2024-12-07 05:38:10.709175] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:37.519 [2024-12-07 05:38:10.709210] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:37.519 [2024-12-07 05:38:10.709288] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:37.519 [2024-12-07 05:38:10.709330] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:37.519 pt2 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:37.520 "name": "raid_bdev1", 00:10:37.520 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:37.520 "strip_size_kb": 0, 00:10:37.520 "state": "configuring", 00:10:37.520 "raid_level": "raid1", 00:10:37.520 "superblock": true, 00:10:37.520 "num_base_bdevs": 4, 00:10:37.520 "num_base_bdevs_discovered": 1, 00:10:37.520 "num_base_bdevs_operational": 3, 00:10:37.520 "base_bdevs_list": [ 00:10:37.520 { 00:10:37.520 "name": null, 00:10:37.520 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:37.520 "is_configured": false, 00:10:37.520 "data_offset": 2048, 00:10:37.520 "data_size": 63488 00:10:37.520 }, 00:10:37.520 { 00:10:37.520 "name": "pt2", 00:10:37.520 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:37.520 "is_configured": true, 00:10:37.520 "data_offset": 2048, 00:10:37.520 "data_size": 63488 00:10:37.520 }, 00:10:37.520 { 00:10:37.520 "name": null, 00:10:37.520 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:37.520 "is_configured": false, 00:10:37.520 "data_offset": 2048, 00:10:37.520 "data_size": 63488 00:10:37.520 }, 00:10:37.520 { 00:10:37.520 "name": null, 00:10:37.520 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:37.520 "is_configured": false, 00:10:37.520 "data_offset": 2048, 00:10:37.520 "data_size": 63488 00:10:37.520 } 00:10:37.520 ] 00:10:37.520 }' 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:37.520 05:38:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.780 [2024-12-07 05:38:11.134014] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:37.780 [2024-12-07 05:38:11.134106] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:37.780 [2024-12-07 05:38:11.134131] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:10:37.780 [2024-12-07 05:38:11.134146] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:37.780 [2024-12-07 05:38:11.134649] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:37.780 [2024-12-07 05:38:11.134683] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:37.780 [2024-12-07 05:38:11.134772] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:37.780 [2024-12-07 05:38:11.134802] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:37.780 pt3 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:37.780 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.040 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.040 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:38.040 "name": "raid_bdev1", 00:10:38.040 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:38.040 "strip_size_kb": 0, 00:10:38.040 "state": "configuring", 00:10:38.040 "raid_level": "raid1", 00:10:38.040 "superblock": true, 00:10:38.040 "num_base_bdevs": 4, 00:10:38.040 "num_base_bdevs_discovered": 2, 00:10:38.040 "num_base_bdevs_operational": 3, 00:10:38.040 "base_bdevs_list": [ 00:10:38.040 { 00:10:38.040 "name": null, 00:10:38.040 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:38.040 "is_configured": false, 00:10:38.040 "data_offset": 2048, 00:10:38.040 "data_size": 63488 00:10:38.040 }, 00:10:38.040 { 00:10:38.040 "name": "pt2", 00:10:38.040 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:38.040 "is_configured": true, 00:10:38.040 "data_offset": 2048, 00:10:38.040 "data_size": 63488 00:10:38.040 }, 00:10:38.040 { 00:10:38.040 "name": "pt3", 00:10:38.040 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:38.040 "is_configured": true, 00:10:38.040 "data_offset": 2048, 00:10:38.040 "data_size": 63488 00:10:38.040 }, 00:10:38.040 { 00:10:38.040 "name": null, 00:10:38.040 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:38.040 "is_configured": false, 00:10:38.040 "data_offset": 2048, 00:10:38.040 "data_size": 63488 00:10:38.040 } 00:10:38.040 ] 00:10:38.040 }' 00:10:38.040 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:38.040 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.300 [2024-12-07 05:38:11.629149] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:38.300 [2024-12-07 05:38:11.629232] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:38.300 [2024-12-07 05:38:11.629257] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:10:38.300 [2024-12-07 05:38:11.629272] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:38.300 [2024-12-07 05:38:11.629782] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:38.300 [2024-12-07 05:38:11.629806] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:38.300 [2024-12-07 05:38:11.629897] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:38.300 [2024-12-07 05:38:11.629923] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:38.300 [2024-12-07 05:38:11.630035] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:38.300 [2024-12-07 05:38:11.630047] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:38.300 [2024-12-07 05:38:11.630331] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:38.300 [2024-12-07 05:38:11.630494] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:38.300 [2024-12-07 05:38:11.630515] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:10:38.300 [2024-12-07 05:38:11.630667] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:38.300 pt4 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:38.300 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:38.301 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:38.301 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:38.301 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:38.301 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.301 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:38.301 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.301 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.301 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.560 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:38.560 "name": "raid_bdev1", 00:10:38.560 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:38.560 "strip_size_kb": 0, 00:10:38.560 "state": "online", 00:10:38.560 "raid_level": "raid1", 00:10:38.560 "superblock": true, 00:10:38.560 "num_base_bdevs": 4, 00:10:38.560 "num_base_bdevs_discovered": 3, 00:10:38.560 "num_base_bdevs_operational": 3, 00:10:38.560 "base_bdevs_list": [ 00:10:38.560 { 00:10:38.560 "name": null, 00:10:38.560 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:38.560 "is_configured": false, 00:10:38.560 "data_offset": 2048, 00:10:38.560 "data_size": 63488 00:10:38.560 }, 00:10:38.560 { 00:10:38.560 "name": "pt2", 00:10:38.560 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:38.560 "is_configured": true, 00:10:38.560 "data_offset": 2048, 00:10:38.560 "data_size": 63488 00:10:38.560 }, 00:10:38.560 { 00:10:38.560 "name": "pt3", 00:10:38.560 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:38.560 "is_configured": true, 00:10:38.560 "data_offset": 2048, 00:10:38.560 "data_size": 63488 00:10:38.560 }, 00:10:38.560 { 00:10:38.560 "name": "pt4", 00:10:38.560 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:38.560 "is_configured": true, 00:10:38.560 "data_offset": 2048, 00:10:38.560 "data_size": 63488 00:10:38.560 } 00:10:38.560 ] 00:10:38.560 }' 00:10:38.560 05:38:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:38.560 05:38:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.820 [2024-12-07 05:38:12.068395] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:38.820 [2024-12-07 05:38:12.068445] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:38.820 [2024-12-07 05:38:12.068538] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:38.820 [2024-12-07 05:38:12.068645] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:38.820 [2024-12-07 05:38:12.068672] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:38.820 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.821 [2024-12-07 05:38:12.140256] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:38.821 [2024-12-07 05:38:12.140314] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:38.821 [2024-12-07 05:38:12.140333] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:10:38.821 [2024-12-07 05:38:12.140342] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:38.821 [2024-12-07 05:38:12.142960] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:38.821 [2024-12-07 05:38:12.142995] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:38.821 [2024-12-07 05:38:12.143080] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:38.821 [2024-12-07 05:38:12.143121] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:38.821 [2024-12-07 05:38:12.143242] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:10:38.821 [2024-12-07 05:38:12.143263] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:38.821 [2024-12-07 05:38:12.143288] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:10:38.821 [2024-12-07 05:38:12.143332] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:38.821 [2024-12-07 05:38:12.143439] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:38.821 pt1 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.821 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.081 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.081 "name": "raid_bdev1", 00:10:39.081 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:39.081 "strip_size_kb": 0, 00:10:39.081 "state": "configuring", 00:10:39.081 "raid_level": "raid1", 00:10:39.081 "superblock": true, 00:10:39.081 "num_base_bdevs": 4, 00:10:39.081 "num_base_bdevs_discovered": 2, 00:10:39.081 "num_base_bdevs_operational": 3, 00:10:39.081 "base_bdevs_list": [ 00:10:39.081 { 00:10:39.081 "name": null, 00:10:39.081 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.081 "is_configured": false, 00:10:39.081 "data_offset": 2048, 00:10:39.081 "data_size": 63488 00:10:39.081 }, 00:10:39.081 { 00:10:39.081 "name": "pt2", 00:10:39.081 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:39.081 "is_configured": true, 00:10:39.081 "data_offset": 2048, 00:10:39.081 "data_size": 63488 00:10:39.081 }, 00:10:39.081 { 00:10:39.081 "name": "pt3", 00:10:39.081 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:39.081 "is_configured": true, 00:10:39.081 "data_offset": 2048, 00:10:39.081 "data_size": 63488 00:10:39.081 }, 00:10:39.081 { 00:10:39.081 "name": null, 00:10:39.081 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:39.081 "is_configured": false, 00:10:39.081 "data_offset": 2048, 00:10:39.081 "data_size": 63488 00:10:39.081 } 00:10:39.081 ] 00:10:39.081 }' 00:10:39.081 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.081 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.340 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:10:39.340 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.340 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.340 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:39.340 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.340 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:10:39.340 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:39.340 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.340 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.340 [2024-12-07 05:38:12.583546] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:39.340 [2024-12-07 05:38:12.583653] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:39.340 [2024-12-07 05:38:12.583680] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:10:39.340 [2024-12-07 05:38:12.583695] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:39.340 [2024-12-07 05:38:12.584226] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:39.340 [2024-12-07 05:38:12.584249] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:39.340 [2024-12-07 05:38:12.584341] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:39.340 [2024-12-07 05:38:12.584374] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:39.340 [2024-12-07 05:38:12.584500] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:10:39.341 [2024-12-07 05:38:12.584513] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:39.341 [2024-12-07 05:38:12.584836] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:10:39.341 [2024-12-07 05:38:12.585020] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:10:39.341 [2024-12-07 05:38:12.585038] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:10:39.341 [2024-12-07 05:38:12.585208] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:39.341 pt4 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.341 "name": "raid_bdev1", 00:10:39.341 "uuid": "529b36b2-ea95-452d-8bf6-d82bb7f31c7e", 00:10:39.341 "strip_size_kb": 0, 00:10:39.341 "state": "online", 00:10:39.341 "raid_level": "raid1", 00:10:39.341 "superblock": true, 00:10:39.341 "num_base_bdevs": 4, 00:10:39.341 "num_base_bdevs_discovered": 3, 00:10:39.341 "num_base_bdevs_operational": 3, 00:10:39.341 "base_bdevs_list": [ 00:10:39.341 { 00:10:39.341 "name": null, 00:10:39.341 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.341 "is_configured": false, 00:10:39.341 "data_offset": 2048, 00:10:39.341 "data_size": 63488 00:10:39.341 }, 00:10:39.341 { 00:10:39.341 "name": "pt2", 00:10:39.341 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:39.341 "is_configured": true, 00:10:39.341 "data_offset": 2048, 00:10:39.341 "data_size": 63488 00:10:39.341 }, 00:10:39.341 { 00:10:39.341 "name": "pt3", 00:10:39.341 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:39.341 "is_configured": true, 00:10:39.341 "data_offset": 2048, 00:10:39.341 "data_size": 63488 00:10:39.341 }, 00:10:39.341 { 00:10:39.341 "name": "pt4", 00:10:39.341 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:39.341 "is_configured": true, 00:10:39.341 "data_offset": 2048, 00:10:39.341 "data_size": 63488 00:10:39.341 } 00:10:39.341 ] 00:10:39.341 }' 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.341 05:38:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.909 [2024-12-07 05:38:13.067001] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 529b36b2-ea95-452d-8bf6-d82bb7f31c7e '!=' 529b36b2-ea95-452d-8bf6-d82bb7f31c7e ']' 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 84932 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 84932 ']' 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 84932 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84932 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:39.909 killing process with pid 84932 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84932' 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 84932 00:10:39.909 [2024-12-07 05:38:13.131141] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:39.909 [2024-12-07 05:38:13.131249] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:39.909 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 84932 00:10:39.909 [2024-12-07 05:38:13.131351] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:39.909 [2024-12-07 05:38:13.131364] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:10:39.909 [2024-12-07 05:38:13.176551] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:40.170 05:38:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:40.170 00:10:40.170 real 0m7.091s 00:10:40.170 user 0m11.893s 00:10:40.170 sys 0m1.536s 00:10:40.170 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:40.170 05:38:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.170 ************************************ 00:10:40.170 END TEST raid_superblock_test 00:10:40.170 ************************************ 00:10:40.170 05:38:13 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:10:40.170 05:38:13 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:40.170 05:38:13 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:40.170 05:38:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:40.170 ************************************ 00:10:40.170 START TEST raid_read_error_test 00:10:40.170 ************************************ 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 read 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.vycFS0xGlJ 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85399 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85399 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 85399 ']' 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:40.170 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:40.170 05:38:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.430 [2024-12-07 05:38:13.562805] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:10:40.430 [2024-12-07 05:38:13.562930] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85399 ] 00:10:40.430 [2024-12-07 05:38:13.716700] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:40.430 [2024-12-07 05:38:13.742505] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:40.430 [2024-12-07 05:38:13.786811] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:40.430 [2024-12-07 05:38:13.786849] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.370 BaseBdev1_malloc 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.370 true 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.370 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.370 [2024-12-07 05:38:14.419647] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:41.370 [2024-12-07 05:38:14.419710] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.371 [2024-12-07 05:38:14.419735] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:41.371 [2024-12-07 05:38:14.419746] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.371 [2024-12-07 05:38:14.421927] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.371 [2024-12-07 05:38:14.421982] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:41.371 BaseBdev1 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.371 BaseBdev2_malloc 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.371 true 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.371 [2024-12-07 05:38:14.460500] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:41.371 [2024-12-07 05:38:14.460557] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.371 [2024-12-07 05:38:14.460577] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:41.371 [2024-12-07 05:38:14.460596] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.371 [2024-12-07 05:38:14.462750] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.371 [2024-12-07 05:38:14.462855] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:41.371 BaseBdev2 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.371 BaseBdev3_malloc 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.371 true 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.371 [2024-12-07 05:38:14.501268] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:41.371 [2024-12-07 05:38:14.501318] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.371 [2024-12-07 05:38:14.501338] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:41.371 [2024-12-07 05:38:14.501348] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.371 [2024-12-07 05:38:14.503456] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.371 [2024-12-07 05:38:14.503497] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:41.371 BaseBdev3 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.371 BaseBdev4_malloc 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.371 true 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.371 [2024-12-07 05:38:14.558942] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:41.371 [2024-12-07 05:38:14.558997] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.371 [2024-12-07 05:38:14.559021] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:41.371 [2024-12-07 05:38:14.559032] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.371 [2024-12-07 05:38:14.561153] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.371 [2024-12-07 05:38:14.561195] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:41.371 BaseBdev4 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.371 [2024-12-07 05:38:14.570959] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:41.371 [2024-12-07 05:38:14.572935] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:41.371 [2024-12-07 05:38:14.573018] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:41.371 [2024-12-07 05:38:14.573071] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:41.371 [2024-12-07 05:38:14.573275] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:41.371 [2024-12-07 05:38:14.573293] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:41.371 [2024-12-07 05:38:14.573560] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:10:41.371 [2024-12-07 05:38:14.573764] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:41.371 [2024-12-07 05:38:14.573781] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:41.371 [2024-12-07 05:38:14.573917] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.371 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.371 "name": "raid_bdev1", 00:10:41.371 "uuid": "e2b9f930-b3d4-4d7a-b2b6-f39303cceb86", 00:10:41.371 "strip_size_kb": 0, 00:10:41.371 "state": "online", 00:10:41.371 "raid_level": "raid1", 00:10:41.371 "superblock": true, 00:10:41.371 "num_base_bdevs": 4, 00:10:41.371 "num_base_bdevs_discovered": 4, 00:10:41.371 "num_base_bdevs_operational": 4, 00:10:41.371 "base_bdevs_list": [ 00:10:41.371 { 00:10:41.371 "name": "BaseBdev1", 00:10:41.371 "uuid": "e255fd64-03bb-5bfb-b262-6047de8fc0b1", 00:10:41.371 "is_configured": true, 00:10:41.371 "data_offset": 2048, 00:10:41.371 "data_size": 63488 00:10:41.371 }, 00:10:41.371 { 00:10:41.371 "name": "BaseBdev2", 00:10:41.371 "uuid": "1d09226a-a8dd-59dd-b7ec-3c4426f5cfb1", 00:10:41.371 "is_configured": true, 00:10:41.371 "data_offset": 2048, 00:10:41.371 "data_size": 63488 00:10:41.371 }, 00:10:41.371 { 00:10:41.371 "name": "BaseBdev3", 00:10:41.371 "uuid": "bc68c5aa-2bb0-54b5-b22c-79dd892e077d", 00:10:41.371 "is_configured": true, 00:10:41.371 "data_offset": 2048, 00:10:41.371 "data_size": 63488 00:10:41.371 }, 00:10:41.371 { 00:10:41.372 "name": "BaseBdev4", 00:10:41.372 "uuid": "134e20c3-b0b0-564e-b1d2-a751f793b957", 00:10:41.372 "is_configured": true, 00:10:41.372 "data_offset": 2048, 00:10:41.372 "data_size": 63488 00:10:41.372 } 00:10:41.372 ] 00:10:41.372 }' 00:10:41.372 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.372 05:38:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.631 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:41.631 05:38:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:41.890 [2024-12-07 05:38:15.066580] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.827 05:38:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.827 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.827 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.827 "name": "raid_bdev1", 00:10:42.827 "uuid": "e2b9f930-b3d4-4d7a-b2b6-f39303cceb86", 00:10:42.827 "strip_size_kb": 0, 00:10:42.827 "state": "online", 00:10:42.827 "raid_level": "raid1", 00:10:42.827 "superblock": true, 00:10:42.827 "num_base_bdevs": 4, 00:10:42.827 "num_base_bdevs_discovered": 4, 00:10:42.827 "num_base_bdevs_operational": 4, 00:10:42.827 "base_bdevs_list": [ 00:10:42.827 { 00:10:42.827 "name": "BaseBdev1", 00:10:42.827 "uuid": "e255fd64-03bb-5bfb-b262-6047de8fc0b1", 00:10:42.827 "is_configured": true, 00:10:42.827 "data_offset": 2048, 00:10:42.827 "data_size": 63488 00:10:42.827 }, 00:10:42.827 { 00:10:42.827 "name": "BaseBdev2", 00:10:42.827 "uuid": "1d09226a-a8dd-59dd-b7ec-3c4426f5cfb1", 00:10:42.827 "is_configured": true, 00:10:42.827 "data_offset": 2048, 00:10:42.827 "data_size": 63488 00:10:42.827 }, 00:10:42.827 { 00:10:42.827 "name": "BaseBdev3", 00:10:42.827 "uuid": "bc68c5aa-2bb0-54b5-b22c-79dd892e077d", 00:10:42.827 "is_configured": true, 00:10:42.827 "data_offset": 2048, 00:10:42.827 "data_size": 63488 00:10:42.827 }, 00:10:42.827 { 00:10:42.827 "name": "BaseBdev4", 00:10:42.827 "uuid": "134e20c3-b0b0-564e-b1d2-a751f793b957", 00:10:42.827 "is_configured": true, 00:10:42.827 "data_offset": 2048, 00:10:42.827 "data_size": 63488 00:10:42.827 } 00:10:42.827 ] 00:10:42.827 }' 00:10:42.827 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.827 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.088 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:43.088 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.088 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.088 [2024-12-07 05:38:16.439751] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:43.088 [2024-12-07 05:38:16.439861] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:43.088 [2024-12-07 05:38:16.442719] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:43.088 [2024-12-07 05:38:16.442840] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:43.088 [2024-12-07 05:38:16.443001] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:43.088 [2024-12-07 05:38:16.443060] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:43.088 { 00:10:43.088 "results": [ 00:10:43.088 { 00:10:43.088 "job": "raid_bdev1", 00:10:43.088 "core_mask": "0x1", 00:10:43.088 "workload": "randrw", 00:10:43.088 "percentage": 50, 00:10:43.088 "status": "finished", 00:10:43.088 "queue_depth": 1, 00:10:43.088 "io_size": 131072, 00:10:43.088 "runtime": 1.374, 00:10:43.088 "iops": 10938.136826783115, 00:10:43.088 "mibps": 1367.2671033478894, 00:10:43.088 "io_failed": 0, 00:10:43.088 "io_timeout": 0, 00:10:43.088 "avg_latency_us": 88.46199775049169, 00:10:43.088 "min_latency_us": 24.705676855895195, 00:10:43.088 "max_latency_us": 1402.2986899563318 00:10:43.088 } 00:10:43.088 ], 00:10:43.088 "core_count": 1 00:10:43.088 } 00:10:43.088 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.088 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85399 00:10:43.088 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 85399 ']' 00:10:43.088 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 85399 00:10:43.088 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:10:43.348 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:43.348 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85399 00:10:43.348 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:43.348 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:43.348 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85399' 00:10:43.348 killing process with pid 85399 00:10:43.348 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 85399 00:10:43.348 [2024-12-07 05:38:16.478928] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:43.348 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 85399 00:10:43.348 [2024-12-07 05:38:16.515725] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:43.608 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:43.608 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.vycFS0xGlJ 00:10:43.609 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:43.609 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:43.609 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:43.609 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:43.609 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:43.609 05:38:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:43.609 00:10:43.609 real 0m3.272s 00:10:43.609 user 0m4.086s 00:10:43.609 sys 0m0.543s 00:10:43.609 ************************************ 00:10:43.609 END TEST raid_read_error_test 00:10:43.609 ************************************ 00:10:43.609 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:43.609 05:38:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.609 05:38:16 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:10:43.609 05:38:16 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:43.609 05:38:16 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:43.609 05:38:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:43.609 ************************************ 00:10:43.609 START TEST raid_write_error_test 00:10:43.609 ************************************ 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 write 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.MULaZR6bJu 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85533 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85533 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 85533 ']' 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:43.609 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:43.609 05:38:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.609 [2024-12-07 05:38:16.909572] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:10:43.609 [2024-12-07 05:38:16.909792] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85533 ] 00:10:43.872 [2024-12-07 05:38:17.064661] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:43.872 [2024-12-07 05:38:17.091419] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:43.872 [2024-12-07 05:38:17.135263] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:43.872 [2024-12-07 05:38:17.135354] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.447 BaseBdev1_malloc 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.447 true 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.447 [2024-12-07 05:38:17.779807] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:44.447 [2024-12-07 05:38:17.779858] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:44.447 [2024-12-07 05:38:17.779900] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:44.447 [2024-12-07 05:38:17.779909] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:44.447 [2024-12-07 05:38:17.782018] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:44.447 [2024-12-07 05:38:17.782121] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:44.447 BaseBdev1 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.447 BaseBdev2_malloc 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.447 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.709 true 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.709 [2024-12-07 05:38:17.820447] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:44.709 [2024-12-07 05:38:17.820501] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:44.709 [2024-12-07 05:38:17.820518] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:44.709 [2024-12-07 05:38:17.820535] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:44.709 [2024-12-07 05:38:17.822749] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:44.709 [2024-12-07 05:38:17.822787] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:44.709 BaseBdev2 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.709 BaseBdev3_malloc 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.709 true 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.709 [2024-12-07 05:38:17.857174] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:44.709 [2024-12-07 05:38:17.857221] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:44.709 [2024-12-07 05:38:17.857240] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:44.709 [2024-12-07 05:38:17.857248] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:44.709 [2024-12-07 05:38:17.859348] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:44.709 [2024-12-07 05:38:17.859384] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:44.709 BaseBdev3 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.709 BaseBdev4_malloc 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.709 true 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.709 [2024-12-07 05:38:17.905683] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:44.709 [2024-12-07 05:38:17.905729] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:44.709 [2024-12-07 05:38:17.905750] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:44.709 [2024-12-07 05:38:17.905759] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:44.709 [2024-12-07 05:38:17.907852] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:44.709 [2024-12-07 05:38:17.907886] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:44.709 BaseBdev4 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.709 [2024-12-07 05:38:17.917710] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:44.709 [2024-12-07 05:38:17.919599] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:44.709 [2024-12-07 05:38:17.919688] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:44.709 [2024-12-07 05:38:17.919741] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:44.709 [2024-12-07 05:38:17.919947] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:44.709 [2024-12-07 05:38:17.919959] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:44.709 [2024-12-07 05:38:17.920215] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:10:44.709 [2024-12-07 05:38:17.920361] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:44.709 [2024-12-07 05:38:17.920373] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:44.709 [2024-12-07 05:38:17.920500] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.709 "name": "raid_bdev1", 00:10:44.709 "uuid": "eaf68004-4378-4922-8c5a-b47bfb477748", 00:10:44.709 "strip_size_kb": 0, 00:10:44.709 "state": "online", 00:10:44.709 "raid_level": "raid1", 00:10:44.709 "superblock": true, 00:10:44.709 "num_base_bdevs": 4, 00:10:44.709 "num_base_bdevs_discovered": 4, 00:10:44.709 "num_base_bdevs_operational": 4, 00:10:44.709 "base_bdevs_list": [ 00:10:44.709 { 00:10:44.709 "name": "BaseBdev1", 00:10:44.709 "uuid": "59d4dbd8-020a-5b94-bb5e-4ab698a6c843", 00:10:44.709 "is_configured": true, 00:10:44.709 "data_offset": 2048, 00:10:44.709 "data_size": 63488 00:10:44.709 }, 00:10:44.709 { 00:10:44.709 "name": "BaseBdev2", 00:10:44.709 "uuid": "1f511a57-73b0-576c-b93b-ba6199517de5", 00:10:44.709 "is_configured": true, 00:10:44.709 "data_offset": 2048, 00:10:44.709 "data_size": 63488 00:10:44.709 }, 00:10:44.709 { 00:10:44.709 "name": "BaseBdev3", 00:10:44.709 "uuid": "4f56efdb-e805-5a49-b71f-9defdb0a2800", 00:10:44.709 "is_configured": true, 00:10:44.709 "data_offset": 2048, 00:10:44.709 "data_size": 63488 00:10:44.709 }, 00:10:44.709 { 00:10:44.709 "name": "BaseBdev4", 00:10:44.709 "uuid": "bdeade54-fb86-57d0-8476-3b14f836e8e1", 00:10:44.709 "is_configured": true, 00:10:44.709 "data_offset": 2048, 00:10:44.709 "data_size": 63488 00:10:44.709 } 00:10:44.709 ] 00:10:44.709 }' 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.709 05:38:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.280 05:38:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:45.280 05:38:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:45.280 [2024-12-07 05:38:18.493152] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.219 [2024-12-07 05:38:19.407906] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:10:46.219 [2024-12-07 05:38:19.407959] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:46.219 [2024-12-07 05:38:19.408211] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000003090 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.219 "name": "raid_bdev1", 00:10:46.219 "uuid": "eaf68004-4378-4922-8c5a-b47bfb477748", 00:10:46.219 "strip_size_kb": 0, 00:10:46.219 "state": "online", 00:10:46.219 "raid_level": "raid1", 00:10:46.219 "superblock": true, 00:10:46.219 "num_base_bdevs": 4, 00:10:46.219 "num_base_bdevs_discovered": 3, 00:10:46.219 "num_base_bdevs_operational": 3, 00:10:46.219 "base_bdevs_list": [ 00:10:46.219 { 00:10:46.219 "name": null, 00:10:46.219 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.219 "is_configured": false, 00:10:46.219 "data_offset": 0, 00:10:46.219 "data_size": 63488 00:10:46.219 }, 00:10:46.219 { 00:10:46.219 "name": "BaseBdev2", 00:10:46.219 "uuid": "1f511a57-73b0-576c-b93b-ba6199517de5", 00:10:46.219 "is_configured": true, 00:10:46.219 "data_offset": 2048, 00:10:46.219 "data_size": 63488 00:10:46.219 }, 00:10:46.219 { 00:10:46.219 "name": "BaseBdev3", 00:10:46.219 "uuid": "4f56efdb-e805-5a49-b71f-9defdb0a2800", 00:10:46.219 "is_configured": true, 00:10:46.219 "data_offset": 2048, 00:10:46.219 "data_size": 63488 00:10:46.219 }, 00:10:46.219 { 00:10:46.219 "name": "BaseBdev4", 00:10:46.219 "uuid": "bdeade54-fb86-57d0-8476-3b14f836e8e1", 00:10:46.219 "is_configured": true, 00:10:46.219 "data_offset": 2048, 00:10:46.219 "data_size": 63488 00:10:46.219 } 00:10:46.219 ] 00:10:46.219 }' 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.219 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.788 [2024-12-07 05:38:19.854620] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:46.788 [2024-12-07 05:38:19.854666] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:46.788 [2024-12-07 05:38:19.857261] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:46.788 [2024-12-07 05:38:19.857370] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:46.788 [2024-12-07 05:38:19.857498] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:46.788 [2024-12-07 05:38:19.857550] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:46.788 { 00:10:46.788 "results": [ 00:10:46.788 { 00:10:46.788 "job": "raid_bdev1", 00:10:46.788 "core_mask": "0x1", 00:10:46.788 "workload": "randrw", 00:10:46.788 "percentage": 50, 00:10:46.788 "status": "finished", 00:10:46.788 "queue_depth": 1, 00:10:46.788 "io_size": 131072, 00:10:46.788 "runtime": 1.362211, 00:10:46.788 "iops": 11912.251479396364, 00:10:46.788 "mibps": 1489.0314349245455, 00:10:46.788 "io_failed": 0, 00:10:46.788 "io_timeout": 0, 00:10:46.788 "avg_latency_us": 81.2351858444993, 00:10:46.788 "min_latency_us": 24.370305676855896, 00:10:46.788 "max_latency_us": 1516.7720524017468 00:10:46.788 } 00:10:46.788 ], 00:10:46.788 "core_count": 1 00:10:46.788 } 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85533 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 85533 ']' 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 85533 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85533 00:10:46.788 killing process with pid 85533 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85533' 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 85533 00:10:46.788 [2024-12-07 05:38:19.893571] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:46.788 05:38:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 85533 00:10:46.788 [2024-12-07 05:38:19.930032] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:46.788 05:38:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.MULaZR6bJu 00:10:46.788 05:38:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:46.788 05:38:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:46.788 05:38:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:46.788 05:38:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:46.788 05:38:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:46.788 05:38:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:46.788 05:38:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:46.788 00:10:46.788 real 0m3.332s 00:10:46.788 user 0m4.237s 00:10:46.788 sys 0m0.538s 00:10:46.788 ************************************ 00:10:46.788 END TEST raid_write_error_test 00:10:46.788 ************************************ 00:10:46.788 05:38:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:46.788 05:38:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.046 05:38:20 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:10:47.046 05:38:20 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:10:47.046 05:38:20 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:10:47.046 05:38:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:10:47.046 05:38:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:47.046 05:38:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:47.046 ************************************ 00:10:47.046 START TEST raid_rebuild_test 00:10:47.046 ************************************ 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false false true 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:10:47.046 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=85660 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 85660 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 85660 ']' 00:10:47.046 05:38:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:47.047 05:38:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:47.047 05:38:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:47.047 05:38:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:47.047 05:38:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.047 [2024-12-07 05:38:20.310105] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:10:47.047 [2024-12-07 05:38:20.310321] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:10:47.047 Zero copy mechanism will not be used. 00:10:47.047 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85660 ] 00:10:47.305 [2024-12-07 05:38:20.467259] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:47.305 [2024-12-07 05:38:20.496113] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:47.305 [2024-12-07 05:38:20.539336] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:47.305 [2024-12-07 05:38:20.539450] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.875 BaseBdev1_malloc 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.875 [2024-12-07 05:38:21.163641] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:10:47.875 [2024-12-07 05:38:21.163743] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:47.875 [2024-12-07 05:38:21.163786] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:47.875 [2024-12-07 05:38:21.163816] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:47.875 [2024-12-07 05:38:21.165920] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:47.875 [2024-12-07 05:38:21.165993] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:47.875 BaseBdev1 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.875 BaseBdev2_malloc 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.875 [2024-12-07 05:38:21.192392] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:10:47.875 [2024-12-07 05:38:21.192466] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:47.875 [2024-12-07 05:38:21.192489] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:47.875 [2024-12-07 05:38:21.192498] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:47.875 [2024-12-07 05:38:21.194566] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:47.875 [2024-12-07 05:38:21.194610] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:47.875 BaseBdev2 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.875 spare_malloc 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.875 spare_delay 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.875 [2024-12-07 05:38:21.232957] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:10:47.875 [2024-12-07 05:38:21.233008] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:47.875 [2024-12-07 05:38:21.233043] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:47.875 [2024-12-07 05:38:21.233051] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:47.875 [2024-12-07 05:38:21.235117] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:47.875 [2024-12-07 05:38:21.235153] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:10:47.875 spare 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.875 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.134 [2024-12-07 05:38:21.244972] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:48.134 [2024-12-07 05:38:21.246812] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:48.134 [2024-12-07 05:38:21.246913] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:48.134 [2024-12-07 05:38:21.246923] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:48.134 [2024-12-07 05:38:21.247182] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:10:48.134 [2024-12-07 05:38:21.247308] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:48.134 [2024-12-07 05:38:21.247321] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:48.134 [2024-12-07 05:38:21.247433] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:48.134 "name": "raid_bdev1", 00:10:48.134 "uuid": "d426abe8-a34f-4612-a543-2a5f2c81b73c", 00:10:48.134 "strip_size_kb": 0, 00:10:48.134 "state": "online", 00:10:48.134 "raid_level": "raid1", 00:10:48.134 "superblock": false, 00:10:48.134 "num_base_bdevs": 2, 00:10:48.134 "num_base_bdevs_discovered": 2, 00:10:48.134 "num_base_bdevs_operational": 2, 00:10:48.134 "base_bdevs_list": [ 00:10:48.134 { 00:10:48.134 "name": "BaseBdev1", 00:10:48.134 "uuid": "7fac6656-e6f2-54c7-8f71-bb8d4f379c03", 00:10:48.134 "is_configured": true, 00:10:48.134 "data_offset": 0, 00:10:48.134 "data_size": 65536 00:10:48.134 }, 00:10:48.134 { 00:10:48.134 "name": "BaseBdev2", 00:10:48.134 "uuid": "e19b0114-457e-597e-a788-3b063ad4408e", 00:10:48.134 "is_configured": true, 00:10:48.134 "data_offset": 0, 00:10:48.134 "data_size": 65536 00:10:48.134 } 00:10:48.134 ] 00:10:48.134 }' 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:48.134 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.394 [2024-12-07 05:38:21.672547] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:10:48.394 05:38:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:10:48.654 [2024-12-07 05:38:21.939875] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:48.654 /dev/nbd0 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:10:48.654 1+0 records in 00:10:48.654 1+0 records out 00:10:48.654 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00036315 s, 11.3 MB/s 00:10:48.654 05:38:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:48.654 05:38:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:10:48.654 05:38:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:48.654 05:38:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:10:48.654 05:38:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:10:48.654 05:38:22 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:10:48.654 05:38:22 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:10:48.654 05:38:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:10:48.654 05:38:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:10:48.654 05:38:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:10:52.848 65536+0 records in 00:10:52.848 65536+0 records out 00:10:52.848 33554432 bytes (34 MB, 32 MiB) copied, 3.74035 s, 9.0 MB/s 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:10:52.848 [2024-12-07 05:38:25.976710] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.848 [2024-12-07 05:38:25.992793] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:52.848 05:38:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:52.848 05:38:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.848 05:38:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:52.848 05:38:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.848 05:38:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.848 05:38:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.848 05:38:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:52.848 "name": "raid_bdev1", 00:10:52.848 "uuid": "d426abe8-a34f-4612-a543-2a5f2c81b73c", 00:10:52.848 "strip_size_kb": 0, 00:10:52.848 "state": "online", 00:10:52.848 "raid_level": "raid1", 00:10:52.848 "superblock": false, 00:10:52.848 "num_base_bdevs": 2, 00:10:52.848 "num_base_bdevs_discovered": 1, 00:10:52.848 "num_base_bdevs_operational": 1, 00:10:52.848 "base_bdevs_list": [ 00:10:52.848 { 00:10:52.848 "name": null, 00:10:52.849 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.849 "is_configured": false, 00:10:52.849 "data_offset": 0, 00:10:52.849 "data_size": 65536 00:10:52.849 }, 00:10:52.849 { 00:10:52.849 "name": "BaseBdev2", 00:10:52.849 "uuid": "e19b0114-457e-597e-a788-3b063ad4408e", 00:10:52.849 "is_configured": true, 00:10:52.849 "data_offset": 0, 00:10:52.849 "data_size": 65536 00:10:52.849 } 00:10:52.849 ] 00:10:52.849 }' 00:10:52.849 05:38:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:52.849 05:38:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.108 05:38:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:10:53.108 05:38:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.108 05:38:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.108 [2024-12-07 05:38:26.428082] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:53.108 [2024-12-07 05:38:26.433291] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06220 00:10:53.108 05:38:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.108 05:38:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:10:53.108 [2024-12-07 05:38:26.435524] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:54.520 "name": "raid_bdev1", 00:10:54.520 "uuid": "d426abe8-a34f-4612-a543-2a5f2c81b73c", 00:10:54.520 "strip_size_kb": 0, 00:10:54.520 "state": "online", 00:10:54.520 "raid_level": "raid1", 00:10:54.520 "superblock": false, 00:10:54.520 "num_base_bdevs": 2, 00:10:54.520 "num_base_bdevs_discovered": 2, 00:10:54.520 "num_base_bdevs_operational": 2, 00:10:54.520 "process": { 00:10:54.520 "type": "rebuild", 00:10:54.520 "target": "spare", 00:10:54.520 "progress": { 00:10:54.520 "blocks": 20480, 00:10:54.520 "percent": 31 00:10:54.520 } 00:10:54.520 }, 00:10:54.520 "base_bdevs_list": [ 00:10:54.520 { 00:10:54.520 "name": "spare", 00:10:54.520 "uuid": "00cde6cb-9b56-59d9-84e7-1408bce3eb40", 00:10:54.520 "is_configured": true, 00:10:54.520 "data_offset": 0, 00:10:54.520 "data_size": 65536 00:10:54.520 }, 00:10:54.520 { 00:10:54.520 "name": "BaseBdev2", 00:10:54.520 "uuid": "e19b0114-457e-597e-a788-3b063ad4408e", 00:10:54.520 "is_configured": true, 00:10:54.520 "data_offset": 0, 00:10:54.520 "data_size": 65536 00:10:54.520 } 00:10:54.520 ] 00:10:54.520 }' 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.520 [2024-12-07 05:38:27.595330] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:54.520 [2024-12-07 05:38:27.640964] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:10:54.520 [2024-12-07 05:38:27.641038] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:54.520 [2024-12-07 05:38:27.641057] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:54.520 [2024-12-07 05:38:27.641065] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:54.520 "name": "raid_bdev1", 00:10:54.520 "uuid": "d426abe8-a34f-4612-a543-2a5f2c81b73c", 00:10:54.520 "strip_size_kb": 0, 00:10:54.520 "state": "online", 00:10:54.520 "raid_level": "raid1", 00:10:54.520 "superblock": false, 00:10:54.520 "num_base_bdevs": 2, 00:10:54.520 "num_base_bdevs_discovered": 1, 00:10:54.520 "num_base_bdevs_operational": 1, 00:10:54.520 "base_bdevs_list": [ 00:10:54.520 { 00:10:54.520 "name": null, 00:10:54.520 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:54.520 "is_configured": false, 00:10:54.520 "data_offset": 0, 00:10:54.520 "data_size": 65536 00:10:54.520 }, 00:10:54.520 { 00:10:54.520 "name": "BaseBdev2", 00:10:54.520 "uuid": "e19b0114-457e-597e-a788-3b063ad4408e", 00:10:54.520 "is_configured": true, 00:10:54.520 "data_offset": 0, 00:10:54.520 "data_size": 65536 00:10:54.520 } 00:10:54.520 ] 00:10:54.520 }' 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:54.520 05:38:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.781 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:10:54.781 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:54.781 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:10:54.781 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:10:54.781 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:54.781 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.781 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:54.781 05:38:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.781 05:38:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.781 05:38:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.040 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:55.040 "name": "raid_bdev1", 00:10:55.040 "uuid": "d426abe8-a34f-4612-a543-2a5f2c81b73c", 00:10:55.040 "strip_size_kb": 0, 00:10:55.040 "state": "online", 00:10:55.040 "raid_level": "raid1", 00:10:55.040 "superblock": false, 00:10:55.040 "num_base_bdevs": 2, 00:10:55.040 "num_base_bdevs_discovered": 1, 00:10:55.040 "num_base_bdevs_operational": 1, 00:10:55.040 "base_bdevs_list": [ 00:10:55.040 { 00:10:55.040 "name": null, 00:10:55.040 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:55.040 "is_configured": false, 00:10:55.040 "data_offset": 0, 00:10:55.040 "data_size": 65536 00:10:55.040 }, 00:10:55.040 { 00:10:55.040 "name": "BaseBdev2", 00:10:55.040 "uuid": "e19b0114-457e-597e-a788-3b063ad4408e", 00:10:55.040 "is_configured": true, 00:10:55.040 "data_offset": 0, 00:10:55.040 "data_size": 65536 00:10:55.040 } 00:10:55.040 ] 00:10:55.040 }' 00:10:55.040 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:55.040 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:10:55.040 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:55.040 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:10:55.040 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:10:55.040 05:38:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.040 05:38:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.040 [2024-12-07 05:38:28.241210] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:55.040 [2024-12-07 05:38:28.246196] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d062f0 00:10:55.040 05:38:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.040 05:38:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:10:55.040 [2024-12-07 05:38:28.248188] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:10:55.978 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:55.978 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:55.978 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:55.979 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:55.979 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:55.979 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.979 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:55.979 05:38:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.979 05:38:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.979 05:38:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.979 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:55.979 "name": "raid_bdev1", 00:10:55.979 "uuid": "d426abe8-a34f-4612-a543-2a5f2c81b73c", 00:10:55.979 "strip_size_kb": 0, 00:10:55.979 "state": "online", 00:10:55.979 "raid_level": "raid1", 00:10:55.979 "superblock": false, 00:10:55.979 "num_base_bdevs": 2, 00:10:55.979 "num_base_bdevs_discovered": 2, 00:10:55.979 "num_base_bdevs_operational": 2, 00:10:55.979 "process": { 00:10:55.979 "type": "rebuild", 00:10:55.979 "target": "spare", 00:10:55.979 "progress": { 00:10:55.979 "blocks": 20480, 00:10:55.979 "percent": 31 00:10:55.979 } 00:10:55.979 }, 00:10:55.979 "base_bdevs_list": [ 00:10:55.979 { 00:10:55.979 "name": "spare", 00:10:55.979 "uuid": "00cde6cb-9b56-59d9-84e7-1408bce3eb40", 00:10:55.979 "is_configured": true, 00:10:55.979 "data_offset": 0, 00:10:55.979 "data_size": 65536 00:10:55.979 }, 00:10:55.979 { 00:10:55.979 "name": "BaseBdev2", 00:10:55.979 "uuid": "e19b0114-457e-597e-a788-3b063ad4408e", 00:10:55.979 "is_configured": true, 00:10:55.979 "data_offset": 0, 00:10:55.979 "data_size": 65536 00:10:55.979 } 00:10:55.979 ] 00:10:55.979 }' 00:10:55.979 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:56.238 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:56.238 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:56.238 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:56.238 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=286 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:56.239 "name": "raid_bdev1", 00:10:56.239 "uuid": "d426abe8-a34f-4612-a543-2a5f2c81b73c", 00:10:56.239 "strip_size_kb": 0, 00:10:56.239 "state": "online", 00:10:56.239 "raid_level": "raid1", 00:10:56.239 "superblock": false, 00:10:56.239 "num_base_bdevs": 2, 00:10:56.239 "num_base_bdevs_discovered": 2, 00:10:56.239 "num_base_bdevs_operational": 2, 00:10:56.239 "process": { 00:10:56.239 "type": "rebuild", 00:10:56.239 "target": "spare", 00:10:56.239 "progress": { 00:10:56.239 "blocks": 22528, 00:10:56.239 "percent": 34 00:10:56.239 } 00:10:56.239 }, 00:10:56.239 "base_bdevs_list": [ 00:10:56.239 { 00:10:56.239 "name": "spare", 00:10:56.239 "uuid": "00cde6cb-9b56-59d9-84e7-1408bce3eb40", 00:10:56.239 "is_configured": true, 00:10:56.239 "data_offset": 0, 00:10:56.239 "data_size": 65536 00:10:56.239 }, 00:10:56.239 { 00:10:56.239 "name": "BaseBdev2", 00:10:56.239 "uuid": "e19b0114-457e-597e-a788-3b063ad4408e", 00:10:56.239 "is_configured": true, 00:10:56.239 "data_offset": 0, 00:10:56.239 "data_size": 65536 00:10:56.239 } 00:10:56.239 ] 00:10:56.239 }' 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:56.239 05:38:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:57.617 "name": "raid_bdev1", 00:10:57.617 "uuid": "d426abe8-a34f-4612-a543-2a5f2c81b73c", 00:10:57.617 "strip_size_kb": 0, 00:10:57.617 "state": "online", 00:10:57.617 "raid_level": "raid1", 00:10:57.617 "superblock": false, 00:10:57.617 "num_base_bdevs": 2, 00:10:57.617 "num_base_bdevs_discovered": 2, 00:10:57.617 "num_base_bdevs_operational": 2, 00:10:57.617 "process": { 00:10:57.617 "type": "rebuild", 00:10:57.617 "target": "spare", 00:10:57.617 "progress": { 00:10:57.617 "blocks": 47104, 00:10:57.617 "percent": 71 00:10:57.617 } 00:10:57.617 }, 00:10:57.617 "base_bdevs_list": [ 00:10:57.617 { 00:10:57.617 "name": "spare", 00:10:57.617 "uuid": "00cde6cb-9b56-59d9-84e7-1408bce3eb40", 00:10:57.617 "is_configured": true, 00:10:57.617 "data_offset": 0, 00:10:57.617 "data_size": 65536 00:10:57.617 }, 00:10:57.617 { 00:10:57.617 "name": "BaseBdev2", 00:10:57.617 "uuid": "e19b0114-457e-597e-a788-3b063ad4408e", 00:10:57.617 "is_configured": true, 00:10:57.617 "data_offset": 0, 00:10:57.617 "data_size": 65536 00:10:57.617 } 00:10:57.617 ] 00:10:57.617 }' 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:57.617 05:38:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:10:58.186 [2024-12-07 05:38:31.460250] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:10:58.186 [2024-12-07 05:38:31.460351] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:10:58.186 [2024-12-07 05:38:31.460397] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:58.445 "name": "raid_bdev1", 00:10:58.445 "uuid": "d426abe8-a34f-4612-a543-2a5f2c81b73c", 00:10:58.445 "strip_size_kb": 0, 00:10:58.445 "state": "online", 00:10:58.445 "raid_level": "raid1", 00:10:58.445 "superblock": false, 00:10:58.445 "num_base_bdevs": 2, 00:10:58.445 "num_base_bdevs_discovered": 2, 00:10:58.445 "num_base_bdevs_operational": 2, 00:10:58.445 "base_bdevs_list": [ 00:10:58.445 { 00:10:58.445 "name": "spare", 00:10:58.445 "uuid": "00cde6cb-9b56-59d9-84e7-1408bce3eb40", 00:10:58.445 "is_configured": true, 00:10:58.445 "data_offset": 0, 00:10:58.445 "data_size": 65536 00:10:58.445 }, 00:10:58.445 { 00:10:58.445 "name": "BaseBdev2", 00:10:58.445 "uuid": "e19b0114-457e-597e-a788-3b063ad4408e", 00:10:58.445 "is_configured": true, 00:10:58.445 "data_offset": 0, 00:10:58.445 "data_size": 65536 00:10:58.445 } 00:10:58.445 ] 00:10:58.445 }' 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:58.445 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:10:58.704 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:10:58.704 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:10:58.704 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:58.704 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:10:58.704 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:10:58.704 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:58.704 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:58.705 "name": "raid_bdev1", 00:10:58.705 "uuid": "d426abe8-a34f-4612-a543-2a5f2c81b73c", 00:10:58.705 "strip_size_kb": 0, 00:10:58.705 "state": "online", 00:10:58.705 "raid_level": "raid1", 00:10:58.705 "superblock": false, 00:10:58.705 "num_base_bdevs": 2, 00:10:58.705 "num_base_bdevs_discovered": 2, 00:10:58.705 "num_base_bdevs_operational": 2, 00:10:58.705 "base_bdevs_list": [ 00:10:58.705 { 00:10:58.705 "name": "spare", 00:10:58.705 "uuid": "00cde6cb-9b56-59d9-84e7-1408bce3eb40", 00:10:58.705 "is_configured": true, 00:10:58.705 "data_offset": 0, 00:10:58.705 "data_size": 65536 00:10:58.705 }, 00:10:58.705 { 00:10:58.705 "name": "BaseBdev2", 00:10:58.705 "uuid": "e19b0114-457e-597e-a788-3b063ad4408e", 00:10:58.705 "is_configured": true, 00:10:58.705 "data_offset": 0, 00:10:58.705 "data_size": 65536 00:10:58.705 } 00:10:58.705 ] 00:10:58.705 }' 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.705 05:38:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.705 05:38:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:58.705 "name": "raid_bdev1", 00:10:58.705 "uuid": "d426abe8-a34f-4612-a543-2a5f2c81b73c", 00:10:58.705 "strip_size_kb": 0, 00:10:58.705 "state": "online", 00:10:58.705 "raid_level": "raid1", 00:10:58.705 "superblock": false, 00:10:58.705 "num_base_bdevs": 2, 00:10:58.705 "num_base_bdevs_discovered": 2, 00:10:58.705 "num_base_bdevs_operational": 2, 00:10:58.705 "base_bdevs_list": [ 00:10:58.705 { 00:10:58.705 "name": "spare", 00:10:58.705 "uuid": "00cde6cb-9b56-59d9-84e7-1408bce3eb40", 00:10:58.705 "is_configured": true, 00:10:58.705 "data_offset": 0, 00:10:58.705 "data_size": 65536 00:10:58.705 }, 00:10:58.705 { 00:10:58.705 "name": "BaseBdev2", 00:10:58.705 "uuid": "e19b0114-457e-597e-a788-3b063ad4408e", 00:10:58.705 "is_configured": true, 00:10:58.705 "data_offset": 0, 00:10:58.705 "data_size": 65536 00:10:58.705 } 00:10:58.705 ] 00:10:58.705 }' 00:10:58.705 05:38:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:58.705 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.274 [2024-12-07 05:38:32.419600] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:59.274 [2024-12-07 05:38:32.419643] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:59.274 [2024-12-07 05:38:32.419733] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:59.274 [2024-12-07 05:38:32.419801] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:59.274 [2024-12-07 05:38:32.419813] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:10:59.274 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:10:59.534 /dev/nbd0 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:10:59.534 1+0 records in 00:10:59.534 1+0 records out 00:10:59.534 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000490216 s, 8.4 MB/s 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:10:59.534 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:10:59.793 /dev/nbd1 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:10:59.793 1+0 records in 00:10:59.793 1+0 records out 00:10:59.793 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000253941 s, 16.1 MB/s 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:10:59.793 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:10:59.794 05:38:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:10:59.794 05:38:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:10:59.794 05:38:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:10:59.794 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:10:59.794 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:10:59.794 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:10:59.794 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:10:59.794 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:10:59.794 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:00.053 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:00.053 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:00.053 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:00.053 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:00.053 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:00.053 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:00.053 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:00.053 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:00.053 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:00.053 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 85660 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 85660 ']' 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 85660 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85660 00:11:00.312 killing process with pid 85660 00:11:00.312 Received shutdown signal, test time was about 60.000000 seconds 00:11:00.312 00:11:00.312 Latency(us) 00:11:00.312 [2024-12-07T05:38:33.680Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:00.312 [2024-12-07T05:38:33.680Z] =================================================================================================================== 00:11:00.312 [2024-12-07T05:38:33.680Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85660' 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 85660 00:11:00.312 [2024-12-07 05:38:33.534408] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:00.312 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 85660 00:11:00.312 [2024-12-07 05:38:33.564255] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:11:00.572 ************************************ 00:11:00.572 END TEST raid_rebuild_test 00:11:00.572 ************************************ 00:11:00.572 00:11:00.572 real 0m13.552s 00:11:00.572 user 0m15.882s 00:11:00.572 sys 0m2.807s 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.572 05:38:33 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:11:00.572 05:38:33 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:00.572 05:38:33 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:00.572 05:38:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:00.572 ************************************ 00:11:00.572 START TEST raid_rebuild_test_sb 00:11:00.572 ************************************ 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=86060 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 86060 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 86060 ']' 00:11:00.572 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:00.572 05:38:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.572 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:00.572 Zero copy mechanism will not be used. 00:11:00.572 [2024-12-07 05:38:33.936206] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:11:00.572 [2024-12-07 05:38:33.936324] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86060 ] 00:11:00.832 [2024-12-07 05:38:34.069987] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:00.832 [2024-12-07 05:38:34.095559] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:00.832 [2024-12-07 05:38:34.138662] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:00.832 [2024-12-07 05:38:34.138696] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:01.401 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:01.401 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:11:01.401 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:01.401 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:01.401 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.401 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.661 BaseBdev1_malloc 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.661 [2024-12-07 05:38:34.782393] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:01.661 [2024-12-07 05:38:34.782532] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:01.661 [2024-12-07 05:38:34.782562] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:01.661 [2024-12-07 05:38:34.782575] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:01.661 [2024-12-07 05:38:34.784705] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:01.661 [2024-12-07 05:38:34.784738] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:01.661 BaseBdev1 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.661 BaseBdev2_malloc 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.661 [2024-12-07 05:38:34.811059] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:01.661 [2024-12-07 05:38:34.811113] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:01.661 [2024-12-07 05:38:34.811136] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:01.661 [2024-12-07 05:38:34.811145] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:01.661 [2024-12-07 05:38:34.813264] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:01.661 [2024-12-07 05:38:34.813342] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:01.661 BaseBdev2 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.661 spare_malloc 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.661 spare_delay 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.661 [2024-12-07 05:38:34.851617] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:01.661 [2024-12-07 05:38:34.851673] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:01.661 [2024-12-07 05:38:34.851691] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:01.661 [2024-12-07 05:38:34.851700] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:01.661 [2024-12-07 05:38:34.853817] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:01.661 [2024-12-07 05:38:34.853850] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:01.661 spare 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.661 [2024-12-07 05:38:34.863665] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:01.661 [2024-12-07 05:38:34.865531] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:01.661 [2024-12-07 05:38:34.865695] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:01.661 [2024-12-07 05:38:34.865709] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:01.661 [2024-12-07 05:38:34.865990] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:01.661 [2024-12-07 05:38:34.866128] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:01.661 [2024-12-07 05:38:34.866141] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:01.661 [2024-12-07 05:38:34.866248] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.661 "name": "raid_bdev1", 00:11:01.661 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:01.661 "strip_size_kb": 0, 00:11:01.661 "state": "online", 00:11:01.661 "raid_level": "raid1", 00:11:01.661 "superblock": true, 00:11:01.661 "num_base_bdevs": 2, 00:11:01.661 "num_base_bdevs_discovered": 2, 00:11:01.661 "num_base_bdevs_operational": 2, 00:11:01.661 "base_bdevs_list": [ 00:11:01.661 { 00:11:01.661 "name": "BaseBdev1", 00:11:01.661 "uuid": "c37122ab-30b2-5283-991d-b590f7b333f3", 00:11:01.661 "is_configured": true, 00:11:01.661 "data_offset": 2048, 00:11:01.661 "data_size": 63488 00:11:01.661 }, 00:11:01.661 { 00:11:01.661 "name": "BaseBdev2", 00:11:01.661 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:01.661 "is_configured": true, 00:11:01.661 "data_offset": 2048, 00:11:01.661 "data_size": 63488 00:11:01.661 } 00:11:01.661 ] 00:11:01.661 }' 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.661 05:38:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.229 [2024-12-07 05:38:35.319136] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:02.229 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:02.229 [2024-12-07 05:38:35.582573] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:02.488 /dev/nbd0 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:02.488 1+0 records in 00:11:02.488 1+0 records out 00:11:02.488 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000257496 s, 15.9 MB/s 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:02.488 05:38:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:11:06.715 63488+0 records in 00:11:06.715 63488+0 records out 00:11:06.715 32505856 bytes (33 MB, 31 MiB) copied, 3.81335 s, 8.5 MB/s 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:06.715 [2024-12-07 05:38:39.659136] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.715 [2024-12-07 05:38:39.691200] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.715 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:06.715 "name": "raid_bdev1", 00:11:06.715 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:06.715 "strip_size_kb": 0, 00:11:06.715 "state": "online", 00:11:06.715 "raid_level": "raid1", 00:11:06.715 "superblock": true, 00:11:06.715 "num_base_bdevs": 2, 00:11:06.715 "num_base_bdevs_discovered": 1, 00:11:06.715 "num_base_bdevs_operational": 1, 00:11:06.715 "base_bdevs_list": [ 00:11:06.715 { 00:11:06.715 "name": null, 00:11:06.715 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:06.715 "is_configured": false, 00:11:06.715 "data_offset": 0, 00:11:06.715 "data_size": 63488 00:11:06.715 }, 00:11:06.715 { 00:11:06.715 "name": "BaseBdev2", 00:11:06.715 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:06.715 "is_configured": true, 00:11:06.715 "data_offset": 2048, 00:11:06.715 "data_size": 63488 00:11:06.716 } 00:11:06.716 ] 00:11:06.716 }' 00:11:06.716 05:38:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:06.716 05:38:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.975 05:38:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:06.975 05:38:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.975 05:38:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.975 [2024-12-07 05:38:40.110567] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:06.975 [2024-12-07 05:38:40.115649] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e280 00:11:06.975 05:38:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.975 [2024-12-07 05:38:40.117531] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:06.975 05:38:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:07.914 "name": "raid_bdev1", 00:11:07.914 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:07.914 "strip_size_kb": 0, 00:11:07.914 "state": "online", 00:11:07.914 "raid_level": "raid1", 00:11:07.914 "superblock": true, 00:11:07.914 "num_base_bdevs": 2, 00:11:07.914 "num_base_bdevs_discovered": 2, 00:11:07.914 "num_base_bdevs_operational": 2, 00:11:07.914 "process": { 00:11:07.914 "type": "rebuild", 00:11:07.914 "target": "spare", 00:11:07.914 "progress": { 00:11:07.914 "blocks": 20480, 00:11:07.914 "percent": 32 00:11:07.914 } 00:11:07.914 }, 00:11:07.914 "base_bdevs_list": [ 00:11:07.914 { 00:11:07.914 "name": "spare", 00:11:07.914 "uuid": "0871ef7b-44e0-53a0-bc4b-04a112a8b640", 00:11:07.914 "is_configured": true, 00:11:07.914 "data_offset": 2048, 00:11:07.914 "data_size": 63488 00:11:07.914 }, 00:11:07.914 { 00:11:07.914 "name": "BaseBdev2", 00:11:07.914 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:07.914 "is_configured": true, 00:11:07.914 "data_offset": 2048, 00:11:07.914 "data_size": 63488 00:11:07.914 } 00:11:07.914 ] 00:11:07.914 }' 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.914 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:08.174 [2024-12-07 05:38:41.281779] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:08.174 [2024-12-07 05:38:41.322297] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:08.174 [2024-12-07 05:38:41.322380] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:08.174 [2024-12-07 05:38:41.322400] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:08.174 [2024-12-07 05:38:41.322409] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:08.174 "name": "raid_bdev1", 00:11:08.174 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:08.174 "strip_size_kb": 0, 00:11:08.174 "state": "online", 00:11:08.174 "raid_level": "raid1", 00:11:08.174 "superblock": true, 00:11:08.174 "num_base_bdevs": 2, 00:11:08.174 "num_base_bdevs_discovered": 1, 00:11:08.174 "num_base_bdevs_operational": 1, 00:11:08.174 "base_bdevs_list": [ 00:11:08.174 { 00:11:08.174 "name": null, 00:11:08.174 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:08.174 "is_configured": false, 00:11:08.174 "data_offset": 0, 00:11:08.174 "data_size": 63488 00:11:08.174 }, 00:11:08.174 { 00:11:08.174 "name": "BaseBdev2", 00:11:08.174 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:08.174 "is_configured": true, 00:11:08.174 "data_offset": 2048, 00:11:08.174 "data_size": 63488 00:11:08.174 } 00:11:08.174 ] 00:11:08.174 }' 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:08.174 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:08.434 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:08.434 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:08.434 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:08.434 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:08.434 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:08.434 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.434 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.434 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:08.434 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:08.434 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.693 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:08.693 "name": "raid_bdev1", 00:11:08.693 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:08.693 "strip_size_kb": 0, 00:11:08.693 "state": "online", 00:11:08.693 "raid_level": "raid1", 00:11:08.693 "superblock": true, 00:11:08.693 "num_base_bdevs": 2, 00:11:08.693 "num_base_bdevs_discovered": 1, 00:11:08.693 "num_base_bdevs_operational": 1, 00:11:08.693 "base_bdevs_list": [ 00:11:08.693 { 00:11:08.693 "name": null, 00:11:08.693 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:08.693 "is_configured": false, 00:11:08.693 "data_offset": 0, 00:11:08.693 "data_size": 63488 00:11:08.693 }, 00:11:08.693 { 00:11:08.693 "name": "BaseBdev2", 00:11:08.693 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:08.693 "is_configured": true, 00:11:08.693 "data_offset": 2048, 00:11:08.693 "data_size": 63488 00:11:08.693 } 00:11:08.693 ] 00:11:08.693 }' 00:11:08.693 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:08.693 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:08.693 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:08.693 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:08.693 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:08.693 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.694 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:08.694 [2024-12-07 05:38:41.914556] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:08.694 [2024-12-07 05:38:41.919559] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e350 00:11:08.694 05:38:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.694 05:38:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:08.694 [2024-12-07 05:38:41.921469] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:09.635 "name": "raid_bdev1", 00:11:09.635 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:09.635 "strip_size_kb": 0, 00:11:09.635 "state": "online", 00:11:09.635 "raid_level": "raid1", 00:11:09.635 "superblock": true, 00:11:09.635 "num_base_bdevs": 2, 00:11:09.635 "num_base_bdevs_discovered": 2, 00:11:09.635 "num_base_bdevs_operational": 2, 00:11:09.635 "process": { 00:11:09.635 "type": "rebuild", 00:11:09.635 "target": "spare", 00:11:09.635 "progress": { 00:11:09.635 "blocks": 20480, 00:11:09.635 "percent": 32 00:11:09.635 } 00:11:09.635 }, 00:11:09.635 "base_bdevs_list": [ 00:11:09.635 { 00:11:09.635 "name": "spare", 00:11:09.635 "uuid": "0871ef7b-44e0-53a0-bc4b-04a112a8b640", 00:11:09.635 "is_configured": true, 00:11:09.635 "data_offset": 2048, 00:11:09.635 "data_size": 63488 00:11:09.635 }, 00:11:09.635 { 00:11:09.635 "name": "BaseBdev2", 00:11:09.635 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:09.635 "is_configured": true, 00:11:09.635 "data_offset": 2048, 00:11:09.635 "data_size": 63488 00:11:09.635 } 00:11:09.635 ] 00:11:09.635 }' 00:11:09.635 05:38:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:09.895 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=300 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:09.895 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:09.896 "name": "raid_bdev1", 00:11:09.896 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:09.896 "strip_size_kb": 0, 00:11:09.896 "state": "online", 00:11:09.896 "raid_level": "raid1", 00:11:09.896 "superblock": true, 00:11:09.896 "num_base_bdevs": 2, 00:11:09.896 "num_base_bdevs_discovered": 2, 00:11:09.896 "num_base_bdevs_operational": 2, 00:11:09.896 "process": { 00:11:09.896 "type": "rebuild", 00:11:09.896 "target": "spare", 00:11:09.896 "progress": { 00:11:09.896 "blocks": 22528, 00:11:09.896 "percent": 35 00:11:09.896 } 00:11:09.896 }, 00:11:09.896 "base_bdevs_list": [ 00:11:09.896 { 00:11:09.896 "name": "spare", 00:11:09.896 "uuid": "0871ef7b-44e0-53a0-bc4b-04a112a8b640", 00:11:09.896 "is_configured": true, 00:11:09.896 "data_offset": 2048, 00:11:09.896 "data_size": 63488 00:11:09.896 }, 00:11:09.896 { 00:11:09.896 "name": "BaseBdev2", 00:11:09.896 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:09.896 "is_configured": true, 00:11:09.896 "data_offset": 2048, 00:11:09.896 "data_size": 63488 00:11:09.896 } 00:11:09.896 ] 00:11:09.896 }' 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:09.896 05:38:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:10.838 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:10.838 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:10.838 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:10.838 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:10.838 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:10.838 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:10.838 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:10.838 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:10.838 05:38:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.838 05:38:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.124 05:38:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:11.124 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:11.124 "name": "raid_bdev1", 00:11:11.124 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:11.124 "strip_size_kb": 0, 00:11:11.124 "state": "online", 00:11:11.124 "raid_level": "raid1", 00:11:11.124 "superblock": true, 00:11:11.124 "num_base_bdevs": 2, 00:11:11.124 "num_base_bdevs_discovered": 2, 00:11:11.124 "num_base_bdevs_operational": 2, 00:11:11.124 "process": { 00:11:11.124 "type": "rebuild", 00:11:11.124 "target": "spare", 00:11:11.124 "progress": { 00:11:11.124 "blocks": 45056, 00:11:11.124 "percent": 70 00:11:11.124 } 00:11:11.124 }, 00:11:11.124 "base_bdevs_list": [ 00:11:11.124 { 00:11:11.124 "name": "spare", 00:11:11.124 "uuid": "0871ef7b-44e0-53a0-bc4b-04a112a8b640", 00:11:11.124 "is_configured": true, 00:11:11.124 "data_offset": 2048, 00:11:11.124 "data_size": 63488 00:11:11.124 }, 00:11:11.124 { 00:11:11.124 "name": "BaseBdev2", 00:11:11.124 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:11.124 "is_configured": true, 00:11:11.124 "data_offset": 2048, 00:11:11.124 "data_size": 63488 00:11:11.124 } 00:11:11.124 ] 00:11:11.124 }' 00:11:11.124 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:11.124 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:11.124 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:11.124 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:11.124 05:38:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:11.701 [2024-12-07 05:38:45.032956] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:11.701 [2024-12-07 05:38:45.033057] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:11.701 [2024-12-07 05:38:45.033172] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:12.268 "name": "raid_bdev1", 00:11:12.268 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:12.268 "strip_size_kb": 0, 00:11:12.268 "state": "online", 00:11:12.268 "raid_level": "raid1", 00:11:12.268 "superblock": true, 00:11:12.268 "num_base_bdevs": 2, 00:11:12.268 "num_base_bdevs_discovered": 2, 00:11:12.268 "num_base_bdevs_operational": 2, 00:11:12.268 "base_bdevs_list": [ 00:11:12.268 { 00:11:12.268 "name": "spare", 00:11:12.268 "uuid": "0871ef7b-44e0-53a0-bc4b-04a112a8b640", 00:11:12.268 "is_configured": true, 00:11:12.268 "data_offset": 2048, 00:11:12.268 "data_size": 63488 00:11:12.268 }, 00:11:12.268 { 00:11:12.268 "name": "BaseBdev2", 00:11:12.268 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:12.268 "is_configured": true, 00:11:12.268 "data_offset": 2048, 00:11:12.268 "data_size": 63488 00:11:12.268 } 00:11:12.268 ] 00:11:12.268 }' 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:12.268 "name": "raid_bdev1", 00:11:12.268 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:12.268 "strip_size_kb": 0, 00:11:12.268 "state": "online", 00:11:12.268 "raid_level": "raid1", 00:11:12.268 "superblock": true, 00:11:12.268 "num_base_bdevs": 2, 00:11:12.268 "num_base_bdevs_discovered": 2, 00:11:12.268 "num_base_bdevs_operational": 2, 00:11:12.268 "base_bdevs_list": [ 00:11:12.268 { 00:11:12.268 "name": "spare", 00:11:12.268 "uuid": "0871ef7b-44e0-53a0-bc4b-04a112a8b640", 00:11:12.268 "is_configured": true, 00:11:12.268 "data_offset": 2048, 00:11:12.268 "data_size": 63488 00:11:12.268 }, 00:11:12.268 { 00:11:12.268 "name": "BaseBdev2", 00:11:12.268 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:12.268 "is_configured": true, 00:11:12.268 "data_offset": 2048, 00:11:12.268 "data_size": 63488 00:11:12.268 } 00:11:12.268 ] 00:11:12.268 }' 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:12.268 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:12.269 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:12.269 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:12.269 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:12.269 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:12.269 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:12.269 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.269 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:12.269 05:38:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.269 05:38:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.527 05:38:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.527 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:12.527 "name": "raid_bdev1", 00:11:12.527 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:12.527 "strip_size_kb": 0, 00:11:12.527 "state": "online", 00:11:12.527 "raid_level": "raid1", 00:11:12.527 "superblock": true, 00:11:12.527 "num_base_bdevs": 2, 00:11:12.527 "num_base_bdevs_discovered": 2, 00:11:12.527 "num_base_bdevs_operational": 2, 00:11:12.527 "base_bdevs_list": [ 00:11:12.527 { 00:11:12.527 "name": "spare", 00:11:12.527 "uuid": "0871ef7b-44e0-53a0-bc4b-04a112a8b640", 00:11:12.527 "is_configured": true, 00:11:12.527 "data_offset": 2048, 00:11:12.527 "data_size": 63488 00:11:12.527 }, 00:11:12.527 { 00:11:12.527 "name": "BaseBdev2", 00:11:12.527 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:12.527 "is_configured": true, 00:11:12.527 "data_offset": 2048, 00:11:12.527 "data_size": 63488 00:11:12.527 } 00:11:12.527 ] 00:11:12.527 }' 00:11:12.527 05:38:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:12.527 05:38:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.786 [2024-12-07 05:38:46.064256] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:12.786 [2024-12-07 05:38:46.064335] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:12.786 [2024-12-07 05:38:46.064429] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:12.786 [2024-12-07 05:38:46.064504] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:12.786 [2024-12-07 05:38:46.064517] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:12.786 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:13.045 /dev/nbd0 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:13.045 1+0 records in 00:11:13.045 1+0 records out 00:11:13.045 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000348122 s, 11.8 MB/s 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:13.045 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:13.046 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:13.305 /dev/nbd1 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:13.305 1+0 records in 00:11:13.305 1+0 records out 00:11:13.305 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000408448 s, 10.0 MB/s 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:13.305 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:13.564 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:13.564 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:13.564 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:13.564 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:13.564 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:13.564 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:13.564 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:13.564 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:13.564 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:13.564 05:38:46 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.824 [2024-12-07 05:38:47.101496] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:13.824 [2024-12-07 05:38:47.101556] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:13.824 [2024-12-07 05:38:47.101577] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:13.824 [2024-12-07 05:38:47.101590] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:13.824 [2024-12-07 05:38:47.103862] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:13.824 [2024-12-07 05:38:47.103942] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:13.824 [2024-12-07 05:38:47.104033] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:13.824 [2024-12-07 05:38:47.104087] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:13.824 [2024-12-07 05:38:47.104228] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:13.824 spare 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.824 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.082 [2024-12-07 05:38:47.204156] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:11:14.082 [2024-12-07 05:38:47.204307] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:14.082 [2024-12-07 05:38:47.204679] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cae960 00:11:14.082 [2024-12-07 05:38:47.204886] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:11:14.082 [2024-12-07 05:38:47.204904] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:11:14.082 [2024-12-07 05:38:47.205063] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.082 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:14.082 "name": "raid_bdev1", 00:11:14.082 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:14.082 "strip_size_kb": 0, 00:11:14.082 "state": "online", 00:11:14.083 "raid_level": "raid1", 00:11:14.083 "superblock": true, 00:11:14.083 "num_base_bdevs": 2, 00:11:14.083 "num_base_bdevs_discovered": 2, 00:11:14.083 "num_base_bdevs_operational": 2, 00:11:14.083 "base_bdevs_list": [ 00:11:14.083 { 00:11:14.083 "name": "spare", 00:11:14.083 "uuid": "0871ef7b-44e0-53a0-bc4b-04a112a8b640", 00:11:14.083 "is_configured": true, 00:11:14.083 "data_offset": 2048, 00:11:14.083 "data_size": 63488 00:11:14.083 }, 00:11:14.083 { 00:11:14.083 "name": "BaseBdev2", 00:11:14.083 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:14.083 "is_configured": true, 00:11:14.083 "data_offset": 2048, 00:11:14.083 "data_size": 63488 00:11:14.083 } 00:11:14.083 ] 00:11:14.083 }' 00:11:14.083 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:14.083 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.341 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:14.341 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:14.341 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:14.341 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:14.341 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:14.342 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.342 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:14.342 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.342 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.342 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.342 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:14.342 "name": "raid_bdev1", 00:11:14.342 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:14.342 "strip_size_kb": 0, 00:11:14.342 "state": "online", 00:11:14.342 "raid_level": "raid1", 00:11:14.342 "superblock": true, 00:11:14.342 "num_base_bdevs": 2, 00:11:14.342 "num_base_bdevs_discovered": 2, 00:11:14.342 "num_base_bdevs_operational": 2, 00:11:14.342 "base_bdevs_list": [ 00:11:14.342 { 00:11:14.342 "name": "spare", 00:11:14.342 "uuid": "0871ef7b-44e0-53a0-bc4b-04a112a8b640", 00:11:14.342 "is_configured": true, 00:11:14.342 "data_offset": 2048, 00:11:14.342 "data_size": 63488 00:11:14.342 }, 00:11:14.342 { 00:11:14.342 "name": "BaseBdev2", 00:11:14.342 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:14.342 "is_configured": true, 00:11:14.342 "data_offset": 2048, 00:11:14.342 "data_size": 63488 00:11:14.342 } 00:11:14.342 ] 00:11:14.342 }' 00:11:14.342 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.601 [2024-12-07 05:38:47.820308] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:14.601 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:14.602 "name": "raid_bdev1", 00:11:14.602 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:14.602 "strip_size_kb": 0, 00:11:14.602 "state": "online", 00:11:14.602 "raid_level": "raid1", 00:11:14.602 "superblock": true, 00:11:14.602 "num_base_bdevs": 2, 00:11:14.602 "num_base_bdevs_discovered": 1, 00:11:14.602 "num_base_bdevs_operational": 1, 00:11:14.602 "base_bdevs_list": [ 00:11:14.602 { 00:11:14.602 "name": null, 00:11:14.602 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:14.602 "is_configured": false, 00:11:14.602 "data_offset": 0, 00:11:14.602 "data_size": 63488 00:11:14.602 }, 00:11:14.602 { 00:11:14.602 "name": "BaseBdev2", 00:11:14.602 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:14.602 "is_configured": true, 00:11:14.602 "data_offset": 2048, 00:11:14.602 "data_size": 63488 00:11:14.602 } 00:11:14.602 ] 00:11:14.602 }' 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:14.602 05:38:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.861 05:38:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:14.861 05:38:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.861 05:38:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.861 [2024-12-07 05:38:48.219695] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:14.861 [2024-12-07 05:38:48.220001] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:14.861 [2024-12-07 05:38:48.220064] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:14.861 [2024-12-07 05:38:48.220128] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:14.861 [2024-12-07 05:38:48.224961] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caea30 00:11:14.861 05:38:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.861 05:38:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:11:14.861 [2024-12-07 05:38:48.226954] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:16.240 "name": "raid_bdev1", 00:11:16.240 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:16.240 "strip_size_kb": 0, 00:11:16.240 "state": "online", 00:11:16.240 "raid_level": "raid1", 00:11:16.240 "superblock": true, 00:11:16.240 "num_base_bdevs": 2, 00:11:16.240 "num_base_bdevs_discovered": 2, 00:11:16.240 "num_base_bdevs_operational": 2, 00:11:16.240 "process": { 00:11:16.240 "type": "rebuild", 00:11:16.240 "target": "spare", 00:11:16.240 "progress": { 00:11:16.240 "blocks": 20480, 00:11:16.240 "percent": 32 00:11:16.240 } 00:11:16.240 }, 00:11:16.240 "base_bdevs_list": [ 00:11:16.240 { 00:11:16.240 "name": "spare", 00:11:16.240 "uuid": "0871ef7b-44e0-53a0-bc4b-04a112a8b640", 00:11:16.240 "is_configured": true, 00:11:16.240 "data_offset": 2048, 00:11:16.240 "data_size": 63488 00:11:16.240 }, 00:11:16.240 { 00:11:16.240 "name": "BaseBdev2", 00:11:16.240 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:16.240 "is_configured": true, 00:11:16.240 "data_offset": 2048, 00:11:16.240 "data_size": 63488 00:11:16.240 } 00:11:16.240 ] 00:11:16.240 }' 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.240 [2024-12-07 05:38:49.363479] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:16.240 [2024-12-07 05:38:49.431298] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:16.240 [2024-12-07 05:38:49.431359] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:16.240 [2024-12-07 05:38:49.431378] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:16.240 [2024-12-07 05:38:49.431386] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:16.240 "name": "raid_bdev1", 00:11:16.240 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:16.240 "strip_size_kb": 0, 00:11:16.240 "state": "online", 00:11:16.240 "raid_level": "raid1", 00:11:16.240 "superblock": true, 00:11:16.240 "num_base_bdevs": 2, 00:11:16.240 "num_base_bdevs_discovered": 1, 00:11:16.240 "num_base_bdevs_operational": 1, 00:11:16.240 "base_bdevs_list": [ 00:11:16.240 { 00:11:16.240 "name": null, 00:11:16.240 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:16.240 "is_configured": false, 00:11:16.240 "data_offset": 0, 00:11:16.240 "data_size": 63488 00:11:16.240 }, 00:11:16.240 { 00:11:16.240 "name": "BaseBdev2", 00:11:16.240 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:16.240 "is_configured": true, 00:11:16.240 "data_offset": 2048, 00:11:16.240 "data_size": 63488 00:11:16.240 } 00:11:16.240 ] 00:11:16.240 }' 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:16.240 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.500 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:16.500 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.500 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.500 [2024-12-07 05:38:49.863434] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:16.500 [2024-12-07 05:38:49.863551] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:16.500 [2024-12-07 05:38:49.863602] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:11:16.500 [2024-12-07 05:38:49.863641] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:16.500 [2024-12-07 05:38:49.864132] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:16.500 [2024-12-07 05:38:49.864195] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:16.500 [2024-12-07 05:38:49.864320] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:16.500 [2024-12-07 05:38:49.864360] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:16.500 [2024-12-07 05:38:49.864409] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:16.500 [2024-12-07 05:38:49.864464] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:16.760 [2024-12-07 05:38:49.869329] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:11:16.760 spare 00:11:16.760 05:38:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.760 05:38:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:11:16.760 [2024-12-07 05:38:49.871306] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:17.700 "name": "raid_bdev1", 00:11:17.700 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:17.700 "strip_size_kb": 0, 00:11:17.700 "state": "online", 00:11:17.700 "raid_level": "raid1", 00:11:17.700 "superblock": true, 00:11:17.700 "num_base_bdevs": 2, 00:11:17.700 "num_base_bdevs_discovered": 2, 00:11:17.700 "num_base_bdevs_operational": 2, 00:11:17.700 "process": { 00:11:17.700 "type": "rebuild", 00:11:17.700 "target": "spare", 00:11:17.700 "progress": { 00:11:17.700 "blocks": 20480, 00:11:17.700 "percent": 32 00:11:17.700 } 00:11:17.700 }, 00:11:17.700 "base_bdevs_list": [ 00:11:17.700 { 00:11:17.700 "name": "spare", 00:11:17.700 "uuid": "0871ef7b-44e0-53a0-bc4b-04a112a8b640", 00:11:17.700 "is_configured": true, 00:11:17.700 "data_offset": 2048, 00:11:17.700 "data_size": 63488 00:11:17.700 }, 00:11:17.700 { 00:11:17.700 "name": "BaseBdev2", 00:11:17.700 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:17.700 "is_configured": true, 00:11:17.700 "data_offset": 2048, 00:11:17.700 "data_size": 63488 00:11:17.700 } 00:11:17.700 ] 00:11:17.700 }' 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.700 05:38:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.700 [2024-12-07 05:38:50.991819] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:17.961 [2024-12-07 05:38:51.075599] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:17.961 [2024-12-07 05:38:51.075684] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:17.961 [2024-12-07 05:38:51.075699] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:17.961 [2024-12-07 05:38:51.075709] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:17.961 "name": "raid_bdev1", 00:11:17.961 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:17.961 "strip_size_kb": 0, 00:11:17.961 "state": "online", 00:11:17.961 "raid_level": "raid1", 00:11:17.961 "superblock": true, 00:11:17.961 "num_base_bdevs": 2, 00:11:17.961 "num_base_bdevs_discovered": 1, 00:11:17.961 "num_base_bdevs_operational": 1, 00:11:17.961 "base_bdevs_list": [ 00:11:17.961 { 00:11:17.961 "name": null, 00:11:17.961 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:17.961 "is_configured": false, 00:11:17.961 "data_offset": 0, 00:11:17.961 "data_size": 63488 00:11:17.961 }, 00:11:17.961 { 00:11:17.961 "name": "BaseBdev2", 00:11:17.961 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:17.961 "is_configured": true, 00:11:17.961 "data_offset": 2048, 00:11:17.961 "data_size": 63488 00:11:17.961 } 00:11:17.961 ] 00:11:17.961 }' 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:17.961 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:18.221 "name": "raid_bdev1", 00:11:18.221 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:18.221 "strip_size_kb": 0, 00:11:18.221 "state": "online", 00:11:18.221 "raid_level": "raid1", 00:11:18.221 "superblock": true, 00:11:18.221 "num_base_bdevs": 2, 00:11:18.221 "num_base_bdevs_discovered": 1, 00:11:18.221 "num_base_bdevs_operational": 1, 00:11:18.221 "base_bdevs_list": [ 00:11:18.221 { 00:11:18.221 "name": null, 00:11:18.221 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:18.221 "is_configured": false, 00:11:18.221 "data_offset": 0, 00:11:18.221 "data_size": 63488 00:11:18.221 }, 00:11:18.221 { 00:11:18.221 "name": "BaseBdev2", 00:11:18.221 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:18.221 "is_configured": true, 00:11:18.221 "data_offset": 2048, 00:11:18.221 "data_size": 63488 00:11:18.221 } 00:11:18.221 ] 00:11:18.221 }' 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.221 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.221 [2024-12-07 05:38:51.575642] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:18.221 [2024-12-07 05:38:51.575720] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:18.221 [2024-12-07 05:38:51.575742] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:11:18.221 [2024-12-07 05:38:51.575755] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:18.221 [2024-12-07 05:38:51.576190] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:18.221 [2024-12-07 05:38:51.576212] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:18.222 [2024-12-07 05:38:51.576289] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:11:18.222 [2024-12-07 05:38:51.576321] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:18.222 [2024-12-07 05:38:51.576329] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:18.222 [2024-12-07 05:38:51.576342] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:11:18.222 BaseBdev1 00:11:18.222 05:38:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.222 05:38:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:11:19.600 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:19.600 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:19.600 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:19.600 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:19.600 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:19.600 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:19.600 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:19.600 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:19.600 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:19.600 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:19.600 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.601 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:19.601 05:38:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.601 05:38:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.601 05:38:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.601 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:19.601 "name": "raid_bdev1", 00:11:19.601 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:19.601 "strip_size_kb": 0, 00:11:19.601 "state": "online", 00:11:19.601 "raid_level": "raid1", 00:11:19.601 "superblock": true, 00:11:19.601 "num_base_bdevs": 2, 00:11:19.601 "num_base_bdevs_discovered": 1, 00:11:19.601 "num_base_bdevs_operational": 1, 00:11:19.601 "base_bdevs_list": [ 00:11:19.601 { 00:11:19.601 "name": null, 00:11:19.601 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:19.601 "is_configured": false, 00:11:19.601 "data_offset": 0, 00:11:19.601 "data_size": 63488 00:11:19.601 }, 00:11:19.601 { 00:11:19.601 "name": "BaseBdev2", 00:11:19.601 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:19.601 "is_configured": true, 00:11:19.601 "data_offset": 2048, 00:11:19.601 "data_size": 63488 00:11:19.601 } 00:11:19.601 ] 00:11:19.601 }' 00:11:19.601 05:38:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:19.601 05:38:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.860 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:19.860 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:19.860 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:19.860 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:19.860 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:19.860 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.860 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:19.860 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.860 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.860 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.860 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:19.860 "name": "raid_bdev1", 00:11:19.860 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:19.860 "strip_size_kb": 0, 00:11:19.860 "state": "online", 00:11:19.860 "raid_level": "raid1", 00:11:19.860 "superblock": true, 00:11:19.860 "num_base_bdevs": 2, 00:11:19.860 "num_base_bdevs_discovered": 1, 00:11:19.860 "num_base_bdevs_operational": 1, 00:11:19.860 "base_bdevs_list": [ 00:11:19.860 { 00:11:19.860 "name": null, 00:11:19.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:19.860 "is_configured": false, 00:11:19.861 "data_offset": 0, 00:11:19.861 "data_size": 63488 00:11:19.861 }, 00:11:19.861 { 00:11:19.861 "name": "BaseBdev2", 00:11:19.861 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:19.861 "is_configured": true, 00:11:19.861 "data_offset": 2048, 00:11:19.861 "data_size": 63488 00:11:19.861 } 00:11:19.861 ] 00:11:19.861 }' 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.861 [2024-12-07 05:38:53.172950] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:19.861 [2024-12-07 05:38:53.173115] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:19.861 [2024-12-07 05:38:53.173127] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:19.861 request: 00:11:19.861 { 00:11:19.861 "base_bdev": "BaseBdev1", 00:11:19.861 "raid_bdev": "raid_bdev1", 00:11:19.861 "method": "bdev_raid_add_base_bdev", 00:11:19.861 "req_id": 1 00:11:19.861 } 00:11:19.861 Got JSON-RPC error response 00:11:19.861 response: 00:11:19.861 { 00:11:19.861 "code": -22, 00:11:19.861 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:11:19.861 } 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:11:19.861 05:38:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:21.238 "name": "raid_bdev1", 00:11:21.238 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:21.238 "strip_size_kb": 0, 00:11:21.238 "state": "online", 00:11:21.238 "raid_level": "raid1", 00:11:21.238 "superblock": true, 00:11:21.238 "num_base_bdevs": 2, 00:11:21.238 "num_base_bdevs_discovered": 1, 00:11:21.238 "num_base_bdevs_operational": 1, 00:11:21.238 "base_bdevs_list": [ 00:11:21.238 { 00:11:21.238 "name": null, 00:11:21.238 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.238 "is_configured": false, 00:11:21.238 "data_offset": 0, 00:11:21.238 "data_size": 63488 00:11:21.238 }, 00:11:21.238 { 00:11:21.238 "name": "BaseBdev2", 00:11:21.238 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:21.238 "is_configured": true, 00:11:21.238 "data_offset": 2048, 00:11:21.238 "data_size": 63488 00:11:21.238 } 00:11:21.238 ] 00:11:21.238 }' 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:21.238 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:21.498 "name": "raid_bdev1", 00:11:21.498 "uuid": "cd6e0e04-21af-4c2d-b63e-5c2ef5422c8c", 00:11:21.498 "strip_size_kb": 0, 00:11:21.498 "state": "online", 00:11:21.498 "raid_level": "raid1", 00:11:21.498 "superblock": true, 00:11:21.498 "num_base_bdevs": 2, 00:11:21.498 "num_base_bdevs_discovered": 1, 00:11:21.498 "num_base_bdevs_operational": 1, 00:11:21.498 "base_bdevs_list": [ 00:11:21.498 { 00:11:21.498 "name": null, 00:11:21.498 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.498 "is_configured": false, 00:11:21.498 "data_offset": 0, 00:11:21.498 "data_size": 63488 00:11:21.498 }, 00:11:21.498 { 00:11:21.498 "name": "BaseBdev2", 00:11:21.498 "uuid": "855f508c-435d-5cf3-885d-9ff2137a3254", 00:11:21.498 "is_configured": true, 00:11:21.498 "data_offset": 2048, 00:11:21.498 "data_size": 63488 00:11:21.498 } 00:11:21.498 ] 00:11:21.498 }' 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 86060 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 86060 ']' 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 86060 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86060 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86060' 00:11:21.498 killing process with pid 86060 00:11:21.498 Received shutdown signal, test time was about 60.000000 seconds 00:11:21.498 00:11:21.498 Latency(us) 00:11:21.498 [2024-12-07T05:38:54.866Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:21.498 [2024-12-07T05:38:54.866Z] =================================================================================================================== 00:11:21.498 [2024-12-07T05:38:54.866Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 86060 00:11:21.498 [2024-12-07 05:38:54.807655] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:21.498 [2024-12-07 05:38:54.807799] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:21.498 05:38:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 86060 00:11:21.498 [2024-12-07 05:38:54.807856] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:21.498 [2024-12-07 05:38:54.807866] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:11:21.498 [2024-12-07 05:38:54.837921] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:21.757 05:38:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:11:21.757 00:11:21.757 real 0m21.198s 00:11:21.758 user 0m26.109s 00:11:21.758 sys 0m3.392s 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:21.758 ************************************ 00:11:21.758 END TEST raid_rebuild_test_sb 00:11:21.758 ************************************ 00:11:21.758 05:38:55 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:11:21.758 05:38:55 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:21.758 05:38:55 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:21.758 05:38:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:21.758 ************************************ 00:11:21.758 START TEST raid_rebuild_test_io 00:11:21.758 ************************************ 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false true true 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:21.758 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:22.016 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:22.016 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:22.016 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:22.016 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=86767 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 86767 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 86767 ']' 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:22.017 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:22.017 05:38:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:22.017 [2024-12-07 05:38:55.207847] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:11:22.017 [2024-12-07 05:38:55.208066] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86767 ] 00:11:22.017 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:22.017 Zero copy mechanism will not be used. 00:11:22.017 [2024-12-07 05:38:55.362358] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:22.276 [2024-12-07 05:38:55.387374] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:22.276 [2024-12-07 05:38:55.430246] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:22.276 [2024-12-07 05:38:55.430361] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:22.857 BaseBdev1_malloc 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:22.857 [2024-12-07 05:38:56.049900] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:22.857 [2024-12-07 05:38:56.049962] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:22.857 [2024-12-07 05:38:56.050011] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:22.857 [2024-12-07 05:38:56.050023] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:22.857 [2024-12-07 05:38:56.052121] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:22.857 [2024-12-07 05:38:56.052221] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:22.857 BaseBdev1 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:22.857 BaseBdev2_malloc 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:22.857 [2024-12-07 05:38:56.078570] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:22.857 [2024-12-07 05:38:56.078641] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:22.857 [2024-12-07 05:38:56.078666] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:22.857 [2024-12-07 05:38:56.078675] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:22.857 [2024-12-07 05:38:56.080776] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:22.857 [2024-12-07 05:38:56.080828] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:22.857 BaseBdev2 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:22.857 spare_malloc 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:22.857 spare_delay 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:22.857 [2024-12-07 05:38:56.119177] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:22.857 [2024-12-07 05:38:56.119224] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:22.857 [2024-12-07 05:38:56.119258] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:22.857 [2024-12-07 05:38:56.119266] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:22.857 [2024-12-07 05:38:56.121339] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:22.857 [2024-12-07 05:38:56.121428] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:22.857 spare 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.857 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:22.857 [2024-12-07 05:38:56.131201] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:22.858 [2024-12-07 05:38:56.133084] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:22.858 [2024-12-07 05:38:56.133187] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:22.858 [2024-12-07 05:38:56.133197] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:22.858 [2024-12-07 05:38:56.133460] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:22.858 [2024-12-07 05:38:56.133587] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:22.858 [2024-12-07 05:38:56.133600] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:22.858 [2024-12-07 05:38:56.133795] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:22.858 "name": "raid_bdev1", 00:11:22.858 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:22.858 "strip_size_kb": 0, 00:11:22.858 "state": "online", 00:11:22.858 "raid_level": "raid1", 00:11:22.858 "superblock": false, 00:11:22.858 "num_base_bdevs": 2, 00:11:22.858 "num_base_bdevs_discovered": 2, 00:11:22.858 "num_base_bdevs_operational": 2, 00:11:22.858 "base_bdevs_list": [ 00:11:22.858 { 00:11:22.858 "name": "BaseBdev1", 00:11:22.858 "uuid": "54890898-f0cd-5519-b81f-7096377f6d24", 00:11:22.858 "is_configured": true, 00:11:22.858 "data_offset": 0, 00:11:22.858 "data_size": 65536 00:11:22.858 }, 00:11:22.858 { 00:11:22.858 "name": "BaseBdev2", 00:11:22.858 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:22.858 "is_configured": true, 00:11:22.858 "data_offset": 0, 00:11:22.858 "data_size": 65536 00:11:22.858 } 00:11:22.858 ] 00:11:22.858 }' 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:22.858 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:23.428 [2024-12-07 05:38:56.598752] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:23.428 [2024-12-07 05:38:56.702272] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.428 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:23.429 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.429 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:23.429 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.429 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:23.429 "name": "raid_bdev1", 00:11:23.429 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:23.429 "strip_size_kb": 0, 00:11:23.429 "state": "online", 00:11:23.429 "raid_level": "raid1", 00:11:23.429 "superblock": false, 00:11:23.429 "num_base_bdevs": 2, 00:11:23.429 "num_base_bdevs_discovered": 1, 00:11:23.429 "num_base_bdevs_operational": 1, 00:11:23.429 "base_bdevs_list": [ 00:11:23.429 { 00:11:23.429 "name": null, 00:11:23.429 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.429 "is_configured": false, 00:11:23.429 "data_offset": 0, 00:11:23.429 "data_size": 65536 00:11:23.429 }, 00:11:23.429 { 00:11:23.429 "name": "BaseBdev2", 00:11:23.429 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:23.429 "is_configured": true, 00:11:23.429 "data_offset": 0, 00:11:23.429 "data_size": 65536 00:11:23.429 } 00:11:23.429 ] 00:11:23.429 }' 00:11:23.429 05:38:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:23.429 05:38:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:23.688 [2024-12-07 05:38:56.814956] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:11:23.688 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:23.688 Zero copy mechanism will not be used. 00:11:23.688 Running I/O for 60 seconds... 00:11:23.948 05:38:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:23.948 05:38:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.948 05:38:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:23.948 [2024-12-07 05:38:57.150193] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:23.948 05:38:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.948 05:38:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:23.948 [2024-12-07 05:38:57.195801] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:23.948 [2024-12-07 05:38:57.197885] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:24.207 [2024-12-07 05:38:57.315999] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:24.207 [2024-12-07 05:38:57.316693] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:24.207 [2024-12-07 05:38:57.542612] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:24.207 [2024-12-07 05:38:57.542924] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:24.725 162.00 IOPS, 486.00 MiB/s [2024-12-07T05:38:58.093Z] [2024-12-07 05:38:57.877742] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:24.725 [2024-12-07 05:38:57.997544] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:24.984 "name": "raid_bdev1", 00:11:24.984 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:24.984 "strip_size_kb": 0, 00:11:24.984 "state": "online", 00:11:24.984 "raid_level": "raid1", 00:11:24.984 "superblock": false, 00:11:24.984 "num_base_bdevs": 2, 00:11:24.984 "num_base_bdevs_discovered": 2, 00:11:24.984 "num_base_bdevs_operational": 2, 00:11:24.984 "process": { 00:11:24.984 "type": "rebuild", 00:11:24.984 "target": "spare", 00:11:24.984 "progress": { 00:11:24.984 "blocks": 10240, 00:11:24.984 "percent": 15 00:11:24.984 } 00:11:24.984 }, 00:11:24.984 "base_bdevs_list": [ 00:11:24.984 { 00:11:24.984 "name": "spare", 00:11:24.984 "uuid": "e46c98fa-1d30-51f9-8a29-d8a6407fbe40", 00:11:24.984 "is_configured": true, 00:11:24.984 "data_offset": 0, 00:11:24.984 "data_size": 65536 00:11:24.984 }, 00:11:24.984 { 00:11:24.984 "name": "BaseBdev2", 00:11:24.984 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:24.984 "is_configured": true, 00:11:24.984 "data_offset": 0, 00:11:24.984 "data_size": 65536 00:11:24.984 } 00:11:24.984 ] 00:11:24.984 }' 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.984 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:24.984 [2024-12-07 05:38:58.342922] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:24.984 [2024-12-07 05:38:58.343009] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:25.243 [2024-12-07 05:38:58.447986] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:25.243 [2024-12-07 05:38:58.450402] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:25.243 [2024-12-07 05:38:58.450487] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:25.243 [2024-12-07 05:38:58.450536] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:25.243 [2024-12-07 05:38:58.467885] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.243 "name": "raid_bdev1", 00:11:25.243 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:25.243 "strip_size_kb": 0, 00:11:25.243 "state": "online", 00:11:25.243 "raid_level": "raid1", 00:11:25.243 "superblock": false, 00:11:25.243 "num_base_bdevs": 2, 00:11:25.243 "num_base_bdevs_discovered": 1, 00:11:25.243 "num_base_bdevs_operational": 1, 00:11:25.243 "base_bdevs_list": [ 00:11:25.243 { 00:11:25.243 "name": null, 00:11:25.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.243 "is_configured": false, 00:11:25.243 "data_offset": 0, 00:11:25.243 "data_size": 65536 00:11:25.243 }, 00:11:25.243 { 00:11:25.243 "name": "BaseBdev2", 00:11:25.243 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:25.243 "is_configured": true, 00:11:25.243 "data_offset": 0, 00:11:25.243 "data_size": 65536 00:11:25.243 } 00:11:25.243 ] 00:11:25.243 }' 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.243 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:25.761 163.50 IOPS, 490.50 MiB/s [2024-12-07T05:38:59.129Z] 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:25.761 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:25.761 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:25.761 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:25.762 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:25.762 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.762 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:25.762 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.762 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:25.762 05:38:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.762 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:25.762 "name": "raid_bdev1", 00:11:25.762 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:25.762 "strip_size_kb": 0, 00:11:25.762 "state": "online", 00:11:25.762 "raid_level": "raid1", 00:11:25.762 "superblock": false, 00:11:25.762 "num_base_bdevs": 2, 00:11:25.762 "num_base_bdevs_discovered": 1, 00:11:25.762 "num_base_bdevs_operational": 1, 00:11:25.762 "base_bdevs_list": [ 00:11:25.762 { 00:11:25.762 "name": null, 00:11:25.762 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.762 "is_configured": false, 00:11:25.762 "data_offset": 0, 00:11:25.762 "data_size": 65536 00:11:25.762 }, 00:11:25.762 { 00:11:25.762 "name": "BaseBdev2", 00:11:25.762 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:25.762 "is_configured": true, 00:11:25.762 "data_offset": 0, 00:11:25.762 "data_size": 65536 00:11:25.762 } 00:11:25.762 ] 00:11:25.762 }' 00:11:25.762 05:38:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:25.762 05:38:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:25.762 05:38:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:25.762 05:38:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:25.762 05:38:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:25.762 05:38:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.762 05:38:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:25.762 [2024-12-07 05:38:59.098415] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:25.762 05:38:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.762 05:38:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:26.020 [2024-12-07 05:38:59.136526] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:11:26.020 [2024-12-07 05:38:59.138439] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:26.020 [2024-12-07 05:38:59.256335] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:26.020 [2024-12-07 05:38:59.256864] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:26.279 [2024-12-07 05:38:59.471738] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:26.279 [2024-12-07 05:38:59.472159] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:26.538 158.33 IOPS, 475.00 MiB/s [2024-12-07T05:38:59.906Z] [2024-12-07 05:38:59.829287] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:26.796 [2024-12-07 05:38:59.954687] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:26.796 [2024-12-07 05:38:59.955030] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:26.796 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:26.796 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:26.796 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:26.796 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:26.796 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:26.796 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:26.796 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:26.796 05:39:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.796 05:39:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:26.796 05:39:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:27.054 "name": "raid_bdev1", 00:11:27.054 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:27.054 "strip_size_kb": 0, 00:11:27.054 "state": "online", 00:11:27.054 "raid_level": "raid1", 00:11:27.054 "superblock": false, 00:11:27.054 "num_base_bdevs": 2, 00:11:27.054 "num_base_bdevs_discovered": 2, 00:11:27.054 "num_base_bdevs_operational": 2, 00:11:27.054 "process": { 00:11:27.054 "type": "rebuild", 00:11:27.054 "target": "spare", 00:11:27.054 "progress": { 00:11:27.054 "blocks": 10240, 00:11:27.054 "percent": 15 00:11:27.054 } 00:11:27.054 }, 00:11:27.054 "base_bdevs_list": [ 00:11:27.054 { 00:11:27.054 "name": "spare", 00:11:27.054 "uuid": "e46c98fa-1d30-51f9-8a29-d8a6407fbe40", 00:11:27.054 "is_configured": true, 00:11:27.054 "data_offset": 0, 00:11:27.054 "data_size": 65536 00:11:27.054 }, 00:11:27.054 { 00:11:27.054 "name": "BaseBdev2", 00:11:27.054 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:27.054 "is_configured": true, 00:11:27.054 "data_offset": 0, 00:11:27.054 "data_size": 65536 00:11:27.054 } 00:11:27.054 ] 00:11:27.054 }' 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=317 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:27.054 "name": "raid_bdev1", 00:11:27.054 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:27.054 "strip_size_kb": 0, 00:11:27.054 "state": "online", 00:11:27.054 "raid_level": "raid1", 00:11:27.054 "superblock": false, 00:11:27.054 "num_base_bdevs": 2, 00:11:27.054 "num_base_bdevs_discovered": 2, 00:11:27.054 "num_base_bdevs_operational": 2, 00:11:27.054 "process": { 00:11:27.054 "type": "rebuild", 00:11:27.054 "target": "spare", 00:11:27.054 "progress": { 00:11:27.054 "blocks": 14336, 00:11:27.054 "percent": 21 00:11:27.054 } 00:11:27.054 }, 00:11:27.054 "base_bdevs_list": [ 00:11:27.054 { 00:11:27.054 "name": "spare", 00:11:27.054 "uuid": "e46c98fa-1d30-51f9-8a29-d8a6407fbe40", 00:11:27.054 "is_configured": true, 00:11:27.054 "data_offset": 0, 00:11:27.054 "data_size": 65536 00:11:27.054 }, 00:11:27.054 { 00:11:27.054 "name": "BaseBdev2", 00:11:27.054 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:27.054 "is_configured": true, 00:11:27.054 "data_offset": 0, 00:11:27.054 "data_size": 65536 00:11:27.054 } 00:11:27.054 ] 00:11:27.054 }' 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:27.054 [2024-12-07 05:39:00.396032] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:27.054 05:39:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:27.313 [2024-12-07 05:39:00.650938] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:11:27.571 135.25 IOPS, 405.75 MiB/s [2024-12-07T05:39:00.939Z] [2024-12-07 05:39:00.892768] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:11:28.137 [2024-12-07 05:39:01.327774] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:28.137 "name": "raid_bdev1", 00:11:28.137 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:28.137 "strip_size_kb": 0, 00:11:28.137 "state": "online", 00:11:28.137 "raid_level": "raid1", 00:11:28.137 "superblock": false, 00:11:28.137 "num_base_bdevs": 2, 00:11:28.137 "num_base_bdevs_discovered": 2, 00:11:28.137 "num_base_bdevs_operational": 2, 00:11:28.137 "process": { 00:11:28.137 "type": "rebuild", 00:11:28.137 "target": "spare", 00:11:28.137 "progress": { 00:11:28.137 "blocks": 28672, 00:11:28.137 "percent": 43 00:11:28.137 } 00:11:28.137 }, 00:11:28.137 "base_bdevs_list": [ 00:11:28.137 { 00:11:28.137 "name": "spare", 00:11:28.137 "uuid": "e46c98fa-1d30-51f9-8a29-d8a6407fbe40", 00:11:28.137 "is_configured": true, 00:11:28.137 "data_offset": 0, 00:11:28.137 "data_size": 65536 00:11:28.137 }, 00:11:28.137 { 00:11:28.137 "name": "BaseBdev2", 00:11:28.137 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:28.137 "is_configured": true, 00:11:28.137 "data_offset": 0, 00:11:28.137 "data_size": 65536 00:11:28.137 } 00:11:28.137 ] 00:11:28.137 }' 00:11:28.137 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:28.396 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:28.396 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:28.396 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:28.396 05:39:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:28.396 [2024-12-07 05:39:01.650893] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:11:28.396 [2024-12-07 05:39:01.758672] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:11:29.225 122.20 IOPS, 366.60 MiB/s [2024-12-07T05:39:02.593Z] 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:29.225 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:29.225 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:29.225 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:29.225 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:29.225 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:29.225 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:29.225 05:39:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.225 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:29.225 05:39:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:29.225 05:39:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.485 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:29.485 "name": "raid_bdev1", 00:11:29.485 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:29.485 "strip_size_kb": 0, 00:11:29.485 "state": "online", 00:11:29.485 "raid_level": "raid1", 00:11:29.485 "superblock": false, 00:11:29.485 "num_base_bdevs": 2, 00:11:29.485 "num_base_bdevs_discovered": 2, 00:11:29.485 "num_base_bdevs_operational": 2, 00:11:29.485 "process": { 00:11:29.485 "type": "rebuild", 00:11:29.485 "target": "spare", 00:11:29.485 "progress": { 00:11:29.485 "blocks": 49152, 00:11:29.485 "percent": 75 00:11:29.485 } 00:11:29.485 }, 00:11:29.485 "base_bdevs_list": [ 00:11:29.485 { 00:11:29.485 "name": "spare", 00:11:29.485 "uuid": "e46c98fa-1d30-51f9-8a29-d8a6407fbe40", 00:11:29.485 "is_configured": true, 00:11:29.485 "data_offset": 0, 00:11:29.485 "data_size": 65536 00:11:29.485 }, 00:11:29.485 { 00:11:29.485 "name": "BaseBdev2", 00:11:29.485 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:29.485 "is_configured": true, 00:11:29.485 "data_offset": 0, 00:11:29.485 "data_size": 65536 00:11:29.485 } 00:11:29.485 ] 00:11:29.485 }' 00:11:29.485 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:29.485 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:29.485 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:29.485 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:29.485 05:39:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:29.745 107.33 IOPS, 322.00 MiB/s [2024-12-07T05:39:03.113Z] [2024-12-07 05:39:02.878417] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:11:29.745 [2024-12-07 05:39:02.980625] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:11:30.005 [2024-12-07 05:39:03.308078] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:30.265 [2024-12-07 05:39:03.413259] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:30.265 [2024-12-07 05:39:03.415717] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:30.551 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:30.551 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:30.551 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:30.551 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:30.551 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:30.552 "name": "raid_bdev1", 00:11:30.552 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:30.552 "strip_size_kb": 0, 00:11:30.552 "state": "online", 00:11:30.552 "raid_level": "raid1", 00:11:30.552 "superblock": false, 00:11:30.552 "num_base_bdevs": 2, 00:11:30.552 "num_base_bdevs_discovered": 2, 00:11:30.552 "num_base_bdevs_operational": 2, 00:11:30.552 "base_bdevs_list": [ 00:11:30.552 { 00:11:30.552 "name": "spare", 00:11:30.552 "uuid": "e46c98fa-1d30-51f9-8a29-d8a6407fbe40", 00:11:30.552 "is_configured": true, 00:11:30.552 "data_offset": 0, 00:11:30.552 "data_size": 65536 00:11:30.552 }, 00:11:30.552 { 00:11:30.552 "name": "BaseBdev2", 00:11:30.552 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:30.552 "is_configured": true, 00:11:30.552 "data_offset": 0, 00:11:30.552 "data_size": 65536 00:11:30.552 } 00:11:30.552 ] 00:11:30.552 }' 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:30.552 97.86 IOPS, 293.57 MiB/s [2024-12-07T05:39:03.920Z] 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:30.552 "name": "raid_bdev1", 00:11:30.552 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:30.552 "strip_size_kb": 0, 00:11:30.552 "state": "online", 00:11:30.552 "raid_level": "raid1", 00:11:30.552 "superblock": false, 00:11:30.552 "num_base_bdevs": 2, 00:11:30.552 "num_base_bdevs_discovered": 2, 00:11:30.552 "num_base_bdevs_operational": 2, 00:11:30.552 "base_bdevs_list": [ 00:11:30.552 { 00:11:30.552 "name": "spare", 00:11:30.552 "uuid": "e46c98fa-1d30-51f9-8a29-d8a6407fbe40", 00:11:30.552 "is_configured": true, 00:11:30.552 "data_offset": 0, 00:11:30.552 "data_size": 65536 00:11:30.552 }, 00:11:30.552 { 00:11:30.552 "name": "BaseBdev2", 00:11:30.552 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:30.552 "is_configured": true, 00:11:30.552 "data_offset": 0, 00:11:30.552 "data_size": 65536 00:11:30.552 } 00:11:30.552 ] 00:11:30.552 }' 00:11:30.552 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:30.835 05:39:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:30.835 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.835 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:30.835 "name": "raid_bdev1", 00:11:30.835 "uuid": "0b758c3a-b462-4bc5-9914-5dc017365a5e", 00:11:30.835 "strip_size_kb": 0, 00:11:30.835 "state": "online", 00:11:30.836 "raid_level": "raid1", 00:11:30.836 "superblock": false, 00:11:30.836 "num_base_bdevs": 2, 00:11:30.836 "num_base_bdevs_discovered": 2, 00:11:30.836 "num_base_bdevs_operational": 2, 00:11:30.836 "base_bdevs_list": [ 00:11:30.836 { 00:11:30.836 "name": "spare", 00:11:30.836 "uuid": "e46c98fa-1d30-51f9-8a29-d8a6407fbe40", 00:11:30.836 "is_configured": true, 00:11:30.836 "data_offset": 0, 00:11:30.836 "data_size": 65536 00:11:30.836 }, 00:11:30.836 { 00:11:30.836 "name": "BaseBdev2", 00:11:30.836 "uuid": "1a1a429f-4869-5a1b-a336-ef2c11022b1f", 00:11:30.836 "is_configured": true, 00:11:30.836 "data_offset": 0, 00:11:30.836 "data_size": 65536 00:11:30.836 } 00:11:30.836 ] 00:11:30.836 }' 00:11:30.836 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:30.836 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.096 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:31.096 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.096 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.356 [2024-12-07 05:39:04.465695] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:31.356 [2024-12-07 05:39:04.465797] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:31.356 00:11:31.356 Latency(us) 00:11:31.356 [2024-12-07T05:39:04.724Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:31.356 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:11:31.356 raid_bdev1 : 7.68 91.76 275.29 0.00 0.00 14240.99 289.76 108978.64 00:11:31.356 [2024-12-07T05:39:04.724Z] =================================================================================================================== 00:11:31.356 [2024-12-07T05:39:04.724Z] Total : 91.76 275.29 0.00 0.00 14240.99 289.76 108978.64 00:11:31.356 [2024-12-07 05:39:04.489145] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:31.356 [2024-12-07 05:39:04.489243] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:31.356 [2024-12-07 05:39:04.489332] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:31.356 [2024-12-07 05:39:04.489379] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:31.356 { 00:11:31.356 "results": [ 00:11:31.356 { 00:11:31.356 "job": "raid_bdev1", 00:11:31.356 "core_mask": "0x1", 00:11:31.356 "workload": "randrw", 00:11:31.356 "percentage": 50, 00:11:31.356 "status": "finished", 00:11:31.356 "queue_depth": 2, 00:11:31.356 "io_size": 3145728, 00:11:31.356 "runtime": 7.682823, 00:11:31.356 "iops": 91.76314487526265, 00:11:31.356 "mibps": 275.289434625788, 00:11:31.356 "io_failed": 0, 00:11:31.356 "io_timeout": 0, 00:11:31.356 "avg_latency_us": 14240.98609433553, 00:11:31.356 "min_latency_us": 289.7606986899563, 00:11:31.356 "max_latency_us": 108978.64104803493 00:11:31.356 } 00:11:31.356 ], 00:11:31.356 "core_count": 1 00:11:31.356 } 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:31.356 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:11:31.617 /dev/nbd0 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:31.617 1+0 records in 00:11:31.617 1+0 records out 00:11:31.617 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000258713 s, 15.8 MB/s 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:31.617 05:39:04 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:11:31.877 /dev/nbd1 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:31.877 1+0 records in 00:11:31.877 1+0 records out 00:11:31.877 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00060412 s, 6.8 MB/s 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:31.877 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:31.878 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:32.137 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 86767 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 86767 ']' 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 86767 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86767 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86767' 00:11:32.396 killing process with pid 86767 00:11:32.396 Received shutdown signal, test time was about 8.830335 seconds 00:11:32.396 00:11:32.396 Latency(us) 00:11:32.396 [2024-12-07T05:39:05.764Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:32.396 [2024-12-07T05:39:05.764Z] =================================================================================================================== 00:11:32.396 [2024-12-07T05:39:05.764Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 86767 00:11:32.396 [2024-12-07 05:39:05.630840] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:32.396 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 86767 00:11:32.396 [2024-12-07 05:39:05.656089] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:11:32.656 00:11:32.656 real 0m10.742s 00:11:32.656 user 0m14.056s 00:11:32.656 sys 0m1.368s 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:32.656 ************************************ 00:11:32.656 END TEST raid_rebuild_test_io 00:11:32.656 ************************************ 00:11:32.656 05:39:05 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:11:32.656 05:39:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:32.656 05:39:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:32.656 05:39:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:32.656 ************************************ 00:11:32.656 START TEST raid_rebuild_test_sb_io 00:11:32.656 ************************************ 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true true true 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87129 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87129 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 87129 ']' 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:32.656 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:32.656 05:39:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:32.656 [2024-12-07 05:39:06.021842] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:11:32.656 [2024-12-07 05:39:06.022060] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:11:32.656 Zero copy mechanism will not be used. 00:11:32.656 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87129 ] 00:11:32.915 [2024-12-07 05:39:06.176601] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:32.915 [2024-12-07 05:39:06.204652] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:32.915 [2024-12-07 05:39:06.249457] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:32.915 [2024-12-07 05:39:06.249494] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:33.485 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:33.485 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:11:33.485 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:33.485 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:33.485 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.485 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.747 BaseBdev1_malloc 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.747 [2024-12-07 05:39:06.866180] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:33.747 [2024-12-07 05:39:06.866237] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:33.747 [2024-12-07 05:39:06.866264] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:33.747 [2024-12-07 05:39:06.866276] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:33.747 [2024-12-07 05:39:06.868416] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:33.747 [2024-12-07 05:39:06.868493] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:33.747 BaseBdev1 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.747 BaseBdev2_malloc 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.747 [2024-12-07 05:39:06.894849] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:33.747 [2024-12-07 05:39:06.894947] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:33.747 [2024-12-07 05:39:06.894973] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:33.747 [2024-12-07 05:39:06.894983] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:33.747 [2024-12-07 05:39:06.897111] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:33.747 [2024-12-07 05:39:06.897153] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:33.747 BaseBdev2 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.747 spare_malloc 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.747 spare_delay 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.747 [2024-12-07 05:39:06.935492] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:33.747 [2024-12-07 05:39:06.935540] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:33.747 [2024-12-07 05:39:06.935558] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:33.747 [2024-12-07 05:39:06.935567] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:33.747 [2024-12-07 05:39:06.937711] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:33.747 [2024-12-07 05:39:06.937785] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:33.747 spare 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.747 [2024-12-07 05:39:06.947528] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:33.747 [2024-12-07 05:39:06.949452] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:33.747 [2024-12-07 05:39:06.949613] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:33.747 [2024-12-07 05:39:06.949643] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:33.747 [2024-12-07 05:39:06.949916] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:33.747 [2024-12-07 05:39:06.950055] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:33.747 [2024-12-07 05:39:06.950080] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:33.747 [2024-12-07 05:39:06.950200] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.747 05:39:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.747 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:33.747 "name": "raid_bdev1", 00:11:33.747 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:33.747 "strip_size_kb": 0, 00:11:33.747 "state": "online", 00:11:33.747 "raid_level": "raid1", 00:11:33.747 "superblock": true, 00:11:33.747 "num_base_bdevs": 2, 00:11:33.747 "num_base_bdevs_discovered": 2, 00:11:33.747 "num_base_bdevs_operational": 2, 00:11:33.747 "base_bdevs_list": [ 00:11:33.747 { 00:11:33.747 "name": "BaseBdev1", 00:11:33.747 "uuid": "6d01d685-4421-51ca-9bb8-75ab70dacf3e", 00:11:33.747 "is_configured": true, 00:11:33.747 "data_offset": 2048, 00:11:33.747 "data_size": 63488 00:11:33.747 }, 00:11:33.747 { 00:11:33.747 "name": "BaseBdev2", 00:11:33.747 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:33.747 "is_configured": true, 00:11:33.747 "data_offset": 2048, 00:11:33.748 "data_size": 63488 00:11:33.748 } 00:11:33.748 ] 00:11:33.748 }' 00:11:33.748 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:33.748 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:34.317 [2024-12-07 05:39:07.411006] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:34.317 [2024-12-07 05:39:07.478603] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:34.317 "name": "raid_bdev1", 00:11:34.317 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:34.317 "strip_size_kb": 0, 00:11:34.317 "state": "online", 00:11:34.317 "raid_level": "raid1", 00:11:34.317 "superblock": true, 00:11:34.317 "num_base_bdevs": 2, 00:11:34.317 "num_base_bdevs_discovered": 1, 00:11:34.317 "num_base_bdevs_operational": 1, 00:11:34.317 "base_bdevs_list": [ 00:11:34.317 { 00:11:34.317 "name": null, 00:11:34.317 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:34.317 "is_configured": false, 00:11:34.317 "data_offset": 0, 00:11:34.317 "data_size": 63488 00:11:34.317 }, 00:11:34.317 { 00:11:34.317 "name": "BaseBdev2", 00:11:34.317 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:34.317 "is_configured": true, 00:11:34.317 "data_offset": 2048, 00:11:34.317 "data_size": 63488 00:11:34.317 } 00:11:34.317 ] 00:11:34.317 }' 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:34.317 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:34.317 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:34.317 Zero copy mechanism will not be used. 00:11:34.317 Running I/O for 60 seconds... 00:11:34.317 [2024-12-07 05:39:07.556143] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:11:34.577 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:34.577 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.577 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:34.577 [2024-12-07 05:39:07.927992] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:34.837 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.837 05:39:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:34.837 [2024-12-07 05:39:07.965548] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:34.837 [2024-12-07 05:39:07.967663] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:34.837 [2024-12-07 05:39:08.080615] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:34.837 [2024-12-07 05:39:08.081143] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:35.096 [2024-12-07 05:39:08.316881] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:35.356 [2024-12-07 05:39:08.545943] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:35.356 237.00 IOPS, 711.00 MiB/s [2024-12-07T05:39:08.724Z] [2024-12-07 05:39:08.660718] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:35.356 [2024-12-07 05:39:08.660987] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:35.616 05:39:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:35.616 05:39:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:35.616 05:39:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:35.616 05:39:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:35.616 05:39:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:35.616 05:39:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:35.616 05:39:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:35.616 05:39:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.616 05:39:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:35.616 05:39:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.876 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:35.876 "name": "raid_bdev1", 00:11:35.876 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:35.876 "strip_size_kb": 0, 00:11:35.876 "state": "online", 00:11:35.876 "raid_level": "raid1", 00:11:35.876 "superblock": true, 00:11:35.876 "num_base_bdevs": 2, 00:11:35.876 "num_base_bdevs_discovered": 2, 00:11:35.876 "num_base_bdevs_operational": 2, 00:11:35.876 "process": { 00:11:35.876 "type": "rebuild", 00:11:35.876 "target": "spare", 00:11:35.876 "progress": { 00:11:35.876 "blocks": 12288, 00:11:35.876 "percent": 19 00:11:35.876 } 00:11:35.876 }, 00:11:35.876 "base_bdevs_list": [ 00:11:35.876 { 00:11:35.876 "name": "spare", 00:11:35.876 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:35.876 "is_configured": true, 00:11:35.876 "data_offset": 2048, 00:11:35.876 "data_size": 63488 00:11:35.876 }, 00:11:35.876 { 00:11:35.876 "name": "BaseBdev2", 00:11:35.876 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:35.876 "is_configured": true, 00:11:35.876 "data_offset": 2048, 00:11:35.876 "data_size": 63488 00:11:35.876 } 00:11:35.876 ] 00:11:35.876 }' 00:11:35.876 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:35.876 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:35.876 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:35.876 [2024-12-07 05:39:09.094241] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:35.876 [2024-12-07 05:39:09.094522] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:35.876 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:35.876 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:35.876 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.876 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:35.876 [2024-12-07 05:39:09.110527] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:36.137 [2024-12-07 05:39:09.302038] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:36.137 [2024-12-07 05:39:09.315057] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:36.137 [2024-12-07 05:39:09.315098] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:36.137 [2024-12-07 05:39:09.315111] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:36.137 [2024-12-07 05:39:09.332165] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:36.137 "name": "raid_bdev1", 00:11:36.137 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:36.137 "strip_size_kb": 0, 00:11:36.137 "state": "online", 00:11:36.137 "raid_level": "raid1", 00:11:36.137 "superblock": true, 00:11:36.137 "num_base_bdevs": 2, 00:11:36.137 "num_base_bdevs_discovered": 1, 00:11:36.137 "num_base_bdevs_operational": 1, 00:11:36.137 "base_bdevs_list": [ 00:11:36.137 { 00:11:36.137 "name": null, 00:11:36.137 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:36.137 "is_configured": false, 00:11:36.137 "data_offset": 0, 00:11:36.137 "data_size": 63488 00:11:36.137 }, 00:11:36.137 { 00:11:36.137 "name": "BaseBdev2", 00:11:36.137 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:36.137 "is_configured": true, 00:11:36.137 "data_offset": 2048, 00:11:36.137 "data_size": 63488 00:11:36.137 } 00:11:36.137 ] 00:11:36.137 }' 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:36.137 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:36.657 185.50 IOPS, 556.50 MiB/s [2024-12-07T05:39:10.025Z] 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:36.657 "name": "raid_bdev1", 00:11:36.657 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:36.657 "strip_size_kb": 0, 00:11:36.657 "state": "online", 00:11:36.657 "raid_level": "raid1", 00:11:36.657 "superblock": true, 00:11:36.657 "num_base_bdevs": 2, 00:11:36.657 "num_base_bdevs_discovered": 1, 00:11:36.657 "num_base_bdevs_operational": 1, 00:11:36.657 "base_bdevs_list": [ 00:11:36.657 { 00:11:36.657 "name": null, 00:11:36.657 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:36.657 "is_configured": false, 00:11:36.657 "data_offset": 0, 00:11:36.657 "data_size": 63488 00:11:36.657 }, 00:11:36.657 { 00:11:36.657 "name": "BaseBdev2", 00:11:36.657 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:36.657 "is_configured": true, 00:11:36.657 "data_offset": 2048, 00:11:36.657 "data_size": 63488 00:11:36.657 } 00:11:36.657 ] 00:11:36.657 }' 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:36.657 [2024-12-07 05:39:09.903610] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:36.657 05:39:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:36.657 [2024-12-07 05:39:09.947256] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:11:36.657 [2024-12-07 05:39:09.949293] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:36.917 [2024-12-07 05:39:10.067878] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:36.917 [2024-12-07 05:39:10.068415] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:37.177 [2024-12-07 05:39:10.298507] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:37.177 [2024-12-07 05:39:10.298883] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:37.177 [2024-12-07 05:39:10.532841] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:37.437 192.33 IOPS, 577.00 MiB/s [2024-12-07T05:39:10.805Z] [2024-12-07 05:39:10.645237] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:37.437 [2024-12-07 05:39:10.645598] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:37.699 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:37.699 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:37.699 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:37.699 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:37.699 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:37.699 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:37.699 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:37.699 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.699 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:37.699 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.699 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:37.699 "name": "raid_bdev1", 00:11:37.699 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:37.699 "strip_size_kb": 0, 00:11:37.699 "state": "online", 00:11:37.699 "raid_level": "raid1", 00:11:37.699 "superblock": true, 00:11:37.699 "num_base_bdevs": 2, 00:11:37.699 "num_base_bdevs_discovered": 2, 00:11:37.699 "num_base_bdevs_operational": 2, 00:11:37.699 "process": { 00:11:37.699 "type": "rebuild", 00:11:37.699 "target": "spare", 00:11:37.699 "progress": { 00:11:37.699 "blocks": 12288, 00:11:37.699 "percent": 19 00:11:37.699 } 00:11:37.699 }, 00:11:37.699 "base_bdevs_list": [ 00:11:37.699 { 00:11:37.699 "name": "spare", 00:11:37.699 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:37.699 "is_configured": true, 00:11:37.699 "data_offset": 2048, 00:11:37.700 "data_size": 63488 00:11:37.700 }, 00:11:37.700 { 00:11:37.700 "name": "BaseBdev2", 00:11:37.700 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:37.700 "is_configured": true, 00:11:37.700 "data_offset": 2048, 00:11:37.700 "data_size": 63488 00:11:37.700 } 00:11:37.700 ] 00:11:37.700 }' 00:11:37.700 05:39:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:37.700 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:37.700 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:37.700 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:37.700 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:37.700 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:37.700 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=328 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.962 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:37.962 "name": "raid_bdev1", 00:11:37.962 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:37.962 "strip_size_kb": 0, 00:11:37.962 "state": "online", 00:11:37.962 "raid_level": "raid1", 00:11:37.962 "superblock": true, 00:11:37.962 "num_base_bdevs": 2, 00:11:37.962 "num_base_bdevs_discovered": 2, 00:11:37.962 "num_base_bdevs_operational": 2, 00:11:37.962 "process": { 00:11:37.962 "type": "rebuild", 00:11:37.962 "target": "spare", 00:11:37.962 "progress": { 00:11:37.962 "blocks": 14336, 00:11:37.962 "percent": 22 00:11:37.962 } 00:11:37.962 }, 00:11:37.962 "base_bdevs_list": [ 00:11:37.962 { 00:11:37.962 "name": "spare", 00:11:37.962 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:37.962 "is_configured": true, 00:11:37.962 "data_offset": 2048, 00:11:37.962 "data_size": 63488 00:11:37.962 }, 00:11:37.962 { 00:11:37.962 "name": "BaseBdev2", 00:11:37.962 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:37.962 "is_configured": true, 00:11:37.963 "data_offset": 2048, 00:11:37.963 "data_size": 63488 00:11:37.963 } 00:11:37.963 ] 00:11:37.963 }' 00:11:37.963 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:37.963 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:37.963 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:37.963 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:37.963 05:39:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:38.223 [2024-12-07 05:39:11.479512] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:11:38.223 [2024-12-07 05:39:11.479900] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:11:38.793 160.50 IOPS, 481.50 MiB/s [2024-12-07T05:39:12.161Z] [2024-12-07 05:39:11.892055] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:11:38.793 [2024-12-07 05:39:12.116494] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.053 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:39.053 "name": "raid_bdev1", 00:11:39.053 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:39.053 "strip_size_kb": 0, 00:11:39.053 "state": "online", 00:11:39.053 "raid_level": "raid1", 00:11:39.053 "superblock": true, 00:11:39.053 "num_base_bdevs": 2, 00:11:39.053 "num_base_bdevs_discovered": 2, 00:11:39.053 "num_base_bdevs_operational": 2, 00:11:39.053 "process": { 00:11:39.053 "type": "rebuild", 00:11:39.053 "target": "spare", 00:11:39.053 "progress": { 00:11:39.053 "blocks": 32768, 00:11:39.053 "percent": 51 00:11:39.053 } 00:11:39.053 }, 00:11:39.053 "base_bdevs_list": [ 00:11:39.053 { 00:11:39.054 "name": "spare", 00:11:39.054 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:39.054 "is_configured": true, 00:11:39.054 "data_offset": 2048, 00:11:39.054 "data_size": 63488 00:11:39.054 }, 00:11:39.054 { 00:11:39.054 "name": "BaseBdev2", 00:11:39.054 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:39.054 "is_configured": true, 00:11:39.054 "data_offset": 2048, 00:11:39.054 "data_size": 63488 00:11:39.054 } 00:11:39.054 ] 00:11:39.054 }' 00:11:39.054 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:39.054 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:39.054 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:39.054 [2024-12-07 05:39:12.329465] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:11:39.054 [2024-12-07 05:39:12.329784] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:11:39.054 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:39.054 05:39:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:39.573 138.00 IOPS, 414.00 MiB/s [2024-12-07T05:39:12.941Z] [2024-12-07 05:39:12.780267] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:11:39.833 [2024-12-07 05:39:12.998773] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:11:40.093 [2024-12-07 05:39:13.338483] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:40.093 "name": "raid_bdev1", 00:11:40.093 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:40.093 "strip_size_kb": 0, 00:11:40.093 "state": "online", 00:11:40.093 "raid_level": "raid1", 00:11:40.093 "superblock": true, 00:11:40.093 "num_base_bdevs": 2, 00:11:40.093 "num_base_bdevs_discovered": 2, 00:11:40.093 "num_base_bdevs_operational": 2, 00:11:40.093 "process": { 00:11:40.093 "type": "rebuild", 00:11:40.093 "target": "spare", 00:11:40.093 "progress": { 00:11:40.093 "blocks": 51200, 00:11:40.093 "percent": 80 00:11:40.093 } 00:11:40.093 }, 00:11:40.093 "base_bdevs_list": [ 00:11:40.093 { 00:11:40.093 "name": "spare", 00:11:40.093 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:40.093 "is_configured": true, 00:11:40.093 "data_offset": 2048, 00:11:40.093 "data_size": 63488 00:11:40.093 }, 00:11:40.093 { 00:11:40.093 "name": "BaseBdev2", 00:11:40.093 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:40.093 "is_configured": true, 00:11:40.093 "data_offset": 2048, 00:11:40.093 "data_size": 63488 00:11:40.093 } 00:11:40.093 ] 00:11:40.093 }' 00:11:40.093 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:40.353 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:40.353 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:40.353 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:40.353 05:39:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:40.922 120.00 IOPS, 360.00 MiB/s [2024-12-07T05:39:14.290Z] [2024-12-07 05:39:14.094722] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:40.922 [2024-12-07 05:39:14.194547] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:40.922 [2024-12-07 05:39:14.196975] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:41.181 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:41.181 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:41.181 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:41.181 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:41.181 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:41.181 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:41.181 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.182 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.182 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.182 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.182 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.441 107.71 IOPS, 323.14 MiB/s [2024-12-07T05:39:14.809Z] 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:41.441 "name": "raid_bdev1", 00:11:41.441 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:41.441 "strip_size_kb": 0, 00:11:41.441 "state": "online", 00:11:41.441 "raid_level": "raid1", 00:11:41.441 "superblock": true, 00:11:41.441 "num_base_bdevs": 2, 00:11:41.441 "num_base_bdevs_discovered": 2, 00:11:41.441 "num_base_bdevs_operational": 2, 00:11:41.441 "base_bdevs_list": [ 00:11:41.441 { 00:11:41.441 "name": "spare", 00:11:41.441 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:41.441 "is_configured": true, 00:11:41.441 "data_offset": 2048, 00:11:41.441 "data_size": 63488 00:11:41.441 }, 00:11:41.441 { 00:11:41.441 "name": "BaseBdev2", 00:11:41.441 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:41.441 "is_configured": true, 00:11:41.441 "data_offset": 2048, 00:11:41.441 "data_size": 63488 00:11:41.441 } 00:11:41.441 ] 00:11:41.441 }' 00:11:41.441 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:41.441 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:41.441 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:41.441 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:41.441 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:11:41.441 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:41.442 "name": "raid_bdev1", 00:11:41.442 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:41.442 "strip_size_kb": 0, 00:11:41.442 "state": "online", 00:11:41.442 "raid_level": "raid1", 00:11:41.442 "superblock": true, 00:11:41.442 "num_base_bdevs": 2, 00:11:41.442 "num_base_bdevs_discovered": 2, 00:11:41.442 "num_base_bdevs_operational": 2, 00:11:41.442 "base_bdevs_list": [ 00:11:41.442 { 00:11:41.442 "name": "spare", 00:11:41.442 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:41.442 "is_configured": true, 00:11:41.442 "data_offset": 2048, 00:11:41.442 "data_size": 63488 00:11:41.442 }, 00:11:41.442 { 00:11:41.442 "name": "BaseBdev2", 00:11:41.442 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:41.442 "is_configured": true, 00:11:41.442 "data_offset": 2048, 00:11:41.442 "data_size": 63488 00:11:41.442 } 00:11:41.442 ] 00:11:41.442 }' 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.442 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.702 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:41.702 "name": "raid_bdev1", 00:11:41.702 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:41.702 "strip_size_kb": 0, 00:11:41.702 "state": "online", 00:11:41.702 "raid_level": "raid1", 00:11:41.702 "superblock": true, 00:11:41.702 "num_base_bdevs": 2, 00:11:41.702 "num_base_bdevs_discovered": 2, 00:11:41.702 "num_base_bdevs_operational": 2, 00:11:41.702 "base_bdevs_list": [ 00:11:41.702 { 00:11:41.702 "name": "spare", 00:11:41.702 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:41.702 "is_configured": true, 00:11:41.702 "data_offset": 2048, 00:11:41.702 "data_size": 63488 00:11:41.702 }, 00:11:41.702 { 00:11:41.702 "name": "BaseBdev2", 00:11:41.702 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:41.702 "is_configured": true, 00:11:41.702 "data_offset": 2048, 00:11:41.702 "data_size": 63488 00:11:41.702 } 00:11:41.702 ] 00:11:41.702 }' 00:11:41.702 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:41.702 05:39:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.961 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:41.961 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.961 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.961 [2024-12-07 05:39:15.147811] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:41.961 [2024-12-07 05:39:15.147912] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:41.961 00:11:41.961 Latency(us) 00:11:41.961 [2024-12-07T05:39:15.329Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:41.961 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:11:41.961 raid_bdev1 : 7.62 101.49 304.48 0.00 0.00 13405.23 277.24 113099.68 00:11:41.961 [2024-12-07T05:39:15.329Z] =================================================================================================================== 00:11:41.961 [2024-12-07T05:39:15.329Z] Total : 101.49 304.48 0.00 0.00 13405.23 277.24 113099.68 00:11:41.961 [2024-12-07 05:39:15.163112] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:41.961 [2024-12-07 05:39:15.163167] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:41.961 [2024-12-07 05:39:15.163239] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:41.961 [2024-12-07 05:39:15.163251] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:41.961 { 00:11:41.961 "results": [ 00:11:41.961 { 00:11:41.961 "job": "raid_bdev1", 00:11:41.961 "core_mask": "0x1", 00:11:41.961 "workload": "randrw", 00:11:41.961 "percentage": 50, 00:11:41.961 "status": "finished", 00:11:41.961 "queue_depth": 2, 00:11:41.961 "io_size": 3145728, 00:11:41.961 "runtime": 7.616243, 00:11:41.961 "iops": 101.4936104323352, 00:11:41.961 "mibps": 304.4808312970056, 00:11:41.961 "io_failed": 0, 00:11:41.961 "io_timeout": 0, 00:11:41.961 "avg_latency_us": 13405.227113780033, 00:11:41.961 "min_latency_us": 277.2401746724891, 00:11:41.961 "max_latency_us": 113099.68209606987 00:11:41.961 } 00:11:41.961 ], 00:11:41.961 "core_count": 1 00:11:41.961 } 00:11:41.961 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.961 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:41.962 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:11:42.220 /dev/nbd0 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:42.220 1+0 records in 00:11:42.220 1+0 records out 00:11:42.220 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000326122 s, 12.6 MB/s 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:42.220 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:11:42.480 /dev/nbd1 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:42.480 1+0 records in 00:11:42.480 1+0 records out 00:11:42.480 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000522707 s, 7.8 MB/s 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:42.480 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:42.739 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:42.739 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:42.739 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:42.739 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:42.739 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:42.739 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:42.739 05:39:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:11:42.739 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:42.739 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:42.739 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:42.739 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:42.739 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:42.739 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:11:42.739 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:42.739 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.998 [2024-12-07 05:39:16.227608] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:42.998 [2024-12-07 05:39:16.227675] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:42.998 [2024-12-07 05:39:16.227694] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:11:42.998 [2024-12-07 05:39:16.227705] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:42.998 [2024-12-07 05:39:16.229888] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:42.998 [2024-12-07 05:39:16.229979] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:42.998 [2024-12-07 05:39:16.230073] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:42.998 [2024-12-07 05:39:16.230112] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:42.998 [2024-12-07 05:39:16.230220] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:42.998 spare 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.998 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.998 [2024-12-07 05:39:16.330135] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:11:42.998 [2024-12-07 05:39:16.330161] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:42.998 [2024-12-07 05:39:16.330423] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027720 00:11:42.999 [2024-12-07 05:39:16.330562] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:11:42.999 [2024-12-07 05:39:16.330580] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:11:42.999 [2024-12-07 05:39:16.330740] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.999 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.258 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:43.258 "name": "raid_bdev1", 00:11:43.258 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:43.258 "strip_size_kb": 0, 00:11:43.258 "state": "online", 00:11:43.258 "raid_level": "raid1", 00:11:43.258 "superblock": true, 00:11:43.258 "num_base_bdevs": 2, 00:11:43.258 "num_base_bdevs_discovered": 2, 00:11:43.258 "num_base_bdevs_operational": 2, 00:11:43.258 "base_bdevs_list": [ 00:11:43.258 { 00:11:43.258 "name": "spare", 00:11:43.258 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:43.258 "is_configured": true, 00:11:43.258 "data_offset": 2048, 00:11:43.258 "data_size": 63488 00:11:43.258 }, 00:11:43.258 { 00:11:43.258 "name": "BaseBdev2", 00:11:43.258 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:43.258 "is_configured": true, 00:11:43.258 "data_offset": 2048, 00:11:43.258 "data_size": 63488 00:11:43.258 } 00:11:43.258 ] 00:11:43.258 }' 00:11:43.258 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:43.258 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:43.518 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:43.518 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:43.518 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:43.518 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:43.518 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:43.518 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:43.518 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:43.518 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.519 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:43.519 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.519 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:43.519 "name": "raid_bdev1", 00:11:43.519 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:43.519 "strip_size_kb": 0, 00:11:43.519 "state": "online", 00:11:43.519 "raid_level": "raid1", 00:11:43.519 "superblock": true, 00:11:43.519 "num_base_bdevs": 2, 00:11:43.519 "num_base_bdevs_discovered": 2, 00:11:43.519 "num_base_bdevs_operational": 2, 00:11:43.519 "base_bdevs_list": [ 00:11:43.519 { 00:11:43.519 "name": "spare", 00:11:43.519 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:43.519 "is_configured": true, 00:11:43.519 "data_offset": 2048, 00:11:43.519 "data_size": 63488 00:11:43.519 }, 00:11:43.519 { 00:11:43.519 "name": "BaseBdev2", 00:11:43.519 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:43.519 "is_configured": true, 00:11:43.519 "data_offset": 2048, 00:11:43.519 "data_size": 63488 00:11:43.519 } 00:11:43.519 ] 00:11:43.519 }' 00:11:43.519 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:43.519 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:43.519 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:43.779 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:43.779 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:43.779 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.779 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:43.779 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:43.780 [2024-12-07 05:39:16.942574] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:43.780 "name": "raid_bdev1", 00:11:43.780 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:43.780 "strip_size_kb": 0, 00:11:43.780 "state": "online", 00:11:43.780 "raid_level": "raid1", 00:11:43.780 "superblock": true, 00:11:43.780 "num_base_bdevs": 2, 00:11:43.780 "num_base_bdevs_discovered": 1, 00:11:43.780 "num_base_bdevs_operational": 1, 00:11:43.780 "base_bdevs_list": [ 00:11:43.780 { 00:11:43.780 "name": null, 00:11:43.780 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:43.780 "is_configured": false, 00:11:43.780 "data_offset": 0, 00:11:43.780 "data_size": 63488 00:11:43.780 }, 00:11:43.780 { 00:11:43.780 "name": "BaseBdev2", 00:11:43.780 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:43.780 "is_configured": true, 00:11:43.780 "data_offset": 2048, 00:11:43.780 "data_size": 63488 00:11:43.780 } 00:11:43.780 ] 00:11:43.780 }' 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:43.780 05:39:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.040 05:39:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:44.040 05:39:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:44.040 05:39:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.040 [2024-12-07 05:39:17.369908] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:44.040 [2024-12-07 05:39:17.370176] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:44.040 [2024-12-07 05:39:17.370236] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:44.040 [2024-12-07 05:39:17.370299] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:44.040 [2024-12-07 05:39:17.375526] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000277f0 00:11:44.040 05:39:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:44.040 05:39:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:11:44.040 [2024-12-07 05:39:17.377502] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:45.420 "name": "raid_bdev1", 00:11:45.420 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:45.420 "strip_size_kb": 0, 00:11:45.420 "state": "online", 00:11:45.420 "raid_level": "raid1", 00:11:45.420 "superblock": true, 00:11:45.420 "num_base_bdevs": 2, 00:11:45.420 "num_base_bdevs_discovered": 2, 00:11:45.420 "num_base_bdevs_operational": 2, 00:11:45.420 "process": { 00:11:45.420 "type": "rebuild", 00:11:45.420 "target": "spare", 00:11:45.420 "progress": { 00:11:45.420 "blocks": 20480, 00:11:45.420 "percent": 32 00:11:45.420 } 00:11:45.420 }, 00:11:45.420 "base_bdevs_list": [ 00:11:45.420 { 00:11:45.420 "name": "spare", 00:11:45.420 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:45.420 "is_configured": true, 00:11:45.420 "data_offset": 2048, 00:11:45.420 "data_size": 63488 00:11:45.420 }, 00:11:45.420 { 00:11:45.420 "name": "BaseBdev2", 00:11:45.420 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:45.420 "is_configured": true, 00:11:45.420 "data_offset": 2048, 00:11:45.420 "data_size": 63488 00:11:45.420 } 00:11:45.420 ] 00:11:45.420 }' 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:45.420 [2024-12-07 05:39:18.518000] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:45.420 [2024-12-07 05:39:18.581762] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:45.420 [2024-12-07 05:39:18.581846] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:45.420 [2024-12-07 05:39:18.581868] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:45.420 [2024-12-07 05:39:18.581876] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:45.420 "name": "raid_bdev1", 00:11:45.420 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:45.420 "strip_size_kb": 0, 00:11:45.420 "state": "online", 00:11:45.420 "raid_level": "raid1", 00:11:45.420 "superblock": true, 00:11:45.420 "num_base_bdevs": 2, 00:11:45.420 "num_base_bdevs_discovered": 1, 00:11:45.420 "num_base_bdevs_operational": 1, 00:11:45.420 "base_bdevs_list": [ 00:11:45.420 { 00:11:45.420 "name": null, 00:11:45.420 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:45.420 "is_configured": false, 00:11:45.420 "data_offset": 0, 00:11:45.420 "data_size": 63488 00:11:45.420 }, 00:11:45.420 { 00:11:45.420 "name": "BaseBdev2", 00:11:45.420 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:45.420 "is_configured": true, 00:11:45.420 "data_offset": 2048, 00:11:45.420 "data_size": 63488 00:11:45.420 } 00:11:45.420 ] 00:11:45.420 }' 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:45.420 05:39:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:45.994 05:39:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:45.994 05:39:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.994 05:39:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:45.994 [2024-12-07 05:39:19.066176] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:45.994 [2024-12-07 05:39:19.066329] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:45.994 [2024-12-07 05:39:19.066377] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:45.994 [2024-12-07 05:39:19.066410] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:45.994 [2024-12-07 05:39:19.066903] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:45.994 [2024-12-07 05:39:19.066969] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:45.994 [2024-12-07 05:39:19.067107] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:45.994 [2024-12-07 05:39:19.067149] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:45.994 [2024-12-07 05:39:19.067207] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:45.994 [2024-12-07 05:39:19.067266] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:45.994 [2024-12-07 05:39:19.072554] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000278c0 00:11:45.994 spare 00:11:45.994 05:39:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.994 05:39:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:11:45.994 [2024-12-07 05:39:19.074562] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:46.937 "name": "raid_bdev1", 00:11:46.937 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:46.937 "strip_size_kb": 0, 00:11:46.937 "state": "online", 00:11:46.937 "raid_level": "raid1", 00:11:46.937 "superblock": true, 00:11:46.937 "num_base_bdevs": 2, 00:11:46.937 "num_base_bdevs_discovered": 2, 00:11:46.937 "num_base_bdevs_operational": 2, 00:11:46.937 "process": { 00:11:46.937 "type": "rebuild", 00:11:46.937 "target": "spare", 00:11:46.937 "progress": { 00:11:46.937 "blocks": 20480, 00:11:46.937 "percent": 32 00:11:46.937 } 00:11:46.937 }, 00:11:46.937 "base_bdevs_list": [ 00:11:46.937 { 00:11:46.937 "name": "spare", 00:11:46.937 "uuid": "cb2af49c-05e2-578c-a890-659bf76d8b95", 00:11:46.937 "is_configured": true, 00:11:46.937 "data_offset": 2048, 00:11:46.937 "data_size": 63488 00:11:46.937 }, 00:11:46.937 { 00:11:46.937 "name": "BaseBdev2", 00:11:46.937 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:46.937 "is_configured": true, 00:11:46.937 "data_offset": 2048, 00:11:46.937 "data_size": 63488 00:11:46.937 } 00:11:46.937 ] 00:11:46.937 }' 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:46.937 [2024-12-07 05:39:20.238946] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:46.937 [2024-12-07 05:39:20.278796] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:46.937 [2024-12-07 05:39:20.278859] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:46.937 [2024-12-07 05:39:20.278874] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:46.937 [2024-12-07 05:39:20.278883] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:46.937 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:47.196 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.196 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:47.196 "name": "raid_bdev1", 00:11:47.196 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:47.196 "strip_size_kb": 0, 00:11:47.196 "state": "online", 00:11:47.196 "raid_level": "raid1", 00:11:47.196 "superblock": true, 00:11:47.196 "num_base_bdevs": 2, 00:11:47.196 "num_base_bdevs_discovered": 1, 00:11:47.196 "num_base_bdevs_operational": 1, 00:11:47.196 "base_bdevs_list": [ 00:11:47.196 { 00:11:47.196 "name": null, 00:11:47.196 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:47.196 "is_configured": false, 00:11:47.196 "data_offset": 0, 00:11:47.196 "data_size": 63488 00:11:47.196 }, 00:11:47.196 { 00:11:47.196 "name": "BaseBdev2", 00:11:47.196 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:47.196 "is_configured": true, 00:11:47.196 "data_offset": 2048, 00:11:47.196 "data_size": 63488 00:11:47.196 } 00:11:47.196 ] 00:11:47.196 }' 00:11:47.196 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:47.196 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:47.467 "name": "raid_bdev1", 00:11:47.467 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:47.467 "strip_size_kb": 0, 00:11:47.467 "state": "online", 00:11:47.467 "raid_level": "raid1", 00:11:47.467 "superblock": true, 00:11:47.467 "num_base_bdevs": 2, 00:11:47.467 "num_base_bdevs_discovered": 1, 00:11:47.467 "num_base_bdevs_operational": 1, 00:11:47.467 "base_bdevs_list": [ 00:11:47.467 { 00:11:47.467 "name": null, 00:11:47.467 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:47.467 "is_configured": false, 00:11:47.467 "data_offset": 0, 00:11:47.467 "data_size": 63488 00:11:47.467 }, 00:11:47.467 { 00:11:47.467 "name": "BaseBdev2", 00:11:47.467 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:47.467 "is_configured": true, 00:11:47.467 "data_offset": 2048, 00:11:47.467 "data_size": 63488 00:11:47.467 } 00:11:47.467 ] 00:11:47.467 }' 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:47.467 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:47.726 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:47.726 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:11:47.726 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.726 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:47.726 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.726 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:47.726 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.726 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:47.726 [2024-12-07 05:39:20.866821] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:47.726 [2024-12-07 05:39:20.866882] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:47.726 [2024-12-07 05:39:20.866903] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:11:47.726 [2024-12-07 05:39:20.866914] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:47.726 [2024-12-07 05:39:20.867292] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:47.726 [2024-12-07 05:39:20.867311] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:47.726 [2024-12-07 05:39:20.867382] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:11:47.726 [2024-12-07 05:39:20.867397] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:47.726 [2024-12-07 05:39:20.867407] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:47.726 [2024-12-07 05:39:20.867419] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:11:47.726 BaseBdev1 00:11:47.726 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.726 05:39:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:48.665 "name": "raid_bdev1", 00:11:48.665 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:48.665 "strip_size_kb": 0, 00:11:48.665 "state": "online", 00:11:48.665 "raid_level": "raid1", 00:11:48.665 "superblock": true, 00:11:48.665 "num_base_bdevs": 2, 00:11:48.665 "num_base_bdevs_discovered": 1, 00:11:48.665 "num_base_bdevs_operational": 1, 00:11:48.665 "base_bdevs_list": [ 00:11:48.665 { 00:11:48.665 "name": null, 00:11:48.665 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:48.665 "is_configured": false, 00:11:48.665 "data_offset": 0, 00:11:48.665 "data_size": 63488 00:11:48.665 }, 00:11:48.665 { 00:11:48.665 "name": "BaseBdev2", 00:11:48.665 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:48.665 "is_configured": true, 00:11:48.665 "data_offset": 2048, 00:11:48.665 "data_size": 63488 00:11:48.665 } 00:11:48.665 ] 00:11:48.665 }' 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:48.665 05:39:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.236 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:49.237 "name": "raid_bdev1", 00:11:49.237 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:49.237 "strip_size_kb": 0, 00:11:49.237 "state": "online", 00:11:49.237 "raid_level": "raid1", 00:11:49.237 "superblock": true, 00:11:49.237 "num_base_bdevs": 2, 00:11:49.237 "num_base_bdevs_discovered": 1, 00:11:49.237 "num_base_bdevs_operational": 1, 00:11:49.237 "base_bdevs_list": [ 00:11:49.237 { 00:11:49.237 "name": null, 00:11:49.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.237 "is_configured": false, 00:11:49.237 "data_offset": 0, 00:11:49.237 "data_size": 63488 00:11:49.237 }, 00:11:49.237 { 00:11:49.237 "name": "BaseBdev2", 00:11:49.237 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:49.237 "is_configured": true, 00:11:49.237 "data_offset": 2048, 00:11:49.237 "data_size": 63488 00:11:49.237 } 00:11:49.237 ] 00:11:49.237 }' 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.237 [2024-12-07 05:39:22.492266] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:49.237 [2024-12-07 05:39:22.492429] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:49.237 [2024-12-07 05:39:22.492441] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:49.237 request: 00:11:49.237 { 00:11:49.237 "base_bdev": "BaseBdev1", 00:11:49.237 "raid_bdev": "raid_bdev1", 00:11:49.237 "method": "bdev_raid_add_base_bdev", 00:11:49.237 "req_id": 1 00:11:49.237 } 00:11:49.237 Got JSON-RPC error response 00:11:49.237 response: 00:11:49.237 { 00:11:49.237 "code": -22, 00:11:49.237 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:11:49.237 } 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:11:49.237 05:39:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.176 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.435 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:50.435 "name": "raid_bdev1", 00:11:50.435 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:50.435 "strip_size_kb": 0, 00:11:50.435 "state": "online", 00:11:50.435 "raid_level": "raid1", 00:11:50.435 "superblock": true, 00:11:50.435 "num_base_bdevs": 2, 00:11:50.435 "num_base_bdevs_discovered": 1, 00:11:50.435 "num_base_bdevs_operational": 1, 00:11:50.435 "base_bdevs_list": [ 00:11:50.435 { 00:11:50.435 "name": null, 00:11:50.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.435 "is_configured": false, 00:11:50.435 "data_offset": 0, 00:11:50.435 "data_size": 63488 00:11:50.435 }, 00:11:50.435 { 00:11:50.435 "name": "BaseBdev2", 00:11:50.435 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:50.435 "is_configured": true, 00:11:50.435 "data_offset": 2048, 00:11:50.435 "data_size": 63488 00:11:50.435 } 00:11:50.435 ] 00:11:50.435 }' 00:11:50.435 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:50.435 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.694 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:50.694 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:50.694 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:50.694 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:50.694 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:50.694 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.694 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.694 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.694 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.694 05:39:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.694 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:50.694 "name": "raid_bdev1", 00:11:50.694 "uuid": "ef18a21a-fc3a-4e7d-a2b9-4ef1441820d2", 00:11:50.694 "strip_size_kb": 0, 00:11:50.694 "state": "online", 00:11:50.694 "raid_level": "raid1", 00:11:50.694 "superblock": true, 00:11:50.694 "num_base_bdevs": 2, 00:11:50.694 "num_base_bdevs_discovered": 1, 00:11:50.694 "num_base_bdevs_operational": 1, 00:11:50.694 "base_bdevs_list": [ 00:11:50.694 { 00:11:50.694 "name": null, 00:11:50.694 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.694 "is_configured": false, 00:11:50.695 "data_offset": 0, 00:11:50.695 "data_size": 63488 00:11:50.695 }, 00:11:50.695 { 00:11:50.695 "name": "BaseBdev2", 00:11:50.695 "uuid": "d3181767-67ba-517e-b796-739d8e53618d", 00:11:50.695 "is_configured": true, 00:11:50.695 "data_offset": 2048, 00:11:50.695 "data_size": 63488 00:11:50.695 } 00:11:50.695 ] 00:11:50.695 }' 00:11:50.695 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 87129 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 87129 ']' 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 87129 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87129 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87129' 00:11:50.954 killing process with pid 87129 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 87129 00:11:50.954 Received shutdown signal, test time was about 16.616493 seconds 00:11:50.954 00:11:50.954 Latency(us) 00:11:50.954 [2024-12-07T05:39:24.322Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:50.954 [2024-12-07T05:39:24.322Z] =================================================================================================================== 00:11:50.954 [2024-12-07T05:39:24.322Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:11:50.954 [2024-12-07 05:39:24.142604] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:50.954 [2024-12-07 05:39:24.142771] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:50.954 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 87129 00:11:50.954 [2024-12-07 05:39:24.142832] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:50.954 [2024-12-07 05:39:24.142843] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:11:50.954 [2024-12-07 05:39:24.168179] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:11:51.213 00:11:51.213 real 0m18.438s 00:11:51.213 user 0m24.487s 00:11:51.213 sys 0m2.041s 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:51.213 ************************************ 00:11:51.213 END TEST raid_rebuild_test_sb_io 00:11:51.213 ************************************ 00:11:51.213 05:39:24 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:11:51.213 05:39:24 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:11:51.213 05:39:24 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:51.213 05:39:24 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:51.213 05:39:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:51.213 ************************************ 00:11:51.213 START TEST raid_rebuild_test 00:11:51.213 ************************************ 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false false true 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=87806 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 87806 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 87806 ']' 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:51.213 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:51.213 05:39:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.213 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:51.213 Zero copy mechanism will not be used. 00:11:51.213 [2024-12-07 05:39:24.529948] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:11:51.213 [2024-12-07 05:39:24.530073] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87806 ] 00:11:51.472 [2024-12-07 05:39:24.685550] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:51.472 [2024-12-07 05:39:24.710958] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:51.472 [2024-12-07 05:39:24.754397] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:51.472 [2024-12-07 05:39:24.754455] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:52.049 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:52.049 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.050 BaseBdev1_malloc 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.050 [2024-12-07 05:39:25.362213] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:52.050 [2024-12-07 05:39:25.362270] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:52.050 [2024-12-07 05:39:25.362294] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:52.050 [2024-12-07 05:39:25.362312] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:52.050 [2024-12-07 05:39:25.364424] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:52.050 [2024-12-07 05:39:25.364461] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:52.050 BaseBdev1 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.050 BaseBdev2_malloc 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.050 [2024-12-07 05:39:25.390816] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:52.050 [2024-12-07 05:39:25.390869] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:52.050 [2024-12-07 05:39:25.390892] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:52.050 [2024-12-07 05:39:25.390900] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:52.050 [2024-12-07 05:39:25.392991] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:52.050 [2024-12-07 05:39:25.393033] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:52.050 BaseBdev2 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.050 BaseBdev3_malloc 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.050 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.309 [2024-12-07 05:39:25.419402] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:11:52.309 [2024-12-07 05:39:25.419456] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:52.309 [2024-12-07 05:39:25.419476] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:52.309 [2024-12-07 05:39:25.419486] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:52.309 [2024-12-07 05:39:25.421645] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:52.309 [2024-12-07 05:39:25.421678] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:52.309 BaseBdev3 00:11:52.309 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.309 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:52.309 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:52.309 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.309 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.309 BaseBdev4_malloc 00:11:52.309 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.310 [2024-12-07 05:39:25.460330] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:11:52.310 [2024-12-07 05:39:25.460384] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:52.310 [2024-12-07 05:39:25.460406] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:11:52.310 [2024-12-07 05:39:25.460416] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:52.310 [2024-12-07 05:39:25.462655] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:52.310 [2024-12-07 05:39:25.462702] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:52.310 BaseBdev4 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.310 spare_malloc 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.310 spare_delay 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.310 [2024-12-07 05:39:25.501004] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:52.310 [2024-12-07 05:39:25.501052] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:52.310 [2024-12-07 05:39:25.501071] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:11:52.310 [2024-12-07 05:39:25.501080] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:52.310 [2024-12-07 05:39:25.503331] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:52.310 [2024-12-07 05:39:25.503369] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:52.310 spare 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.310 [2024-12-07 05:39:25.513064] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:52.310 [2024-12-07 05:39:25.515034] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:52.310 [2024-12-07 05:39:25.515104] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:52.310 [2024-12-07 05:39:25.515161] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:52.310 [2024-12-07 05:39:25.515259] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:52.310 [2024-12-07 05:39:25.515271] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:52.310 [2024-12-07 05:39:25.515553] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:52.310 [2024-12-07 05:39:25.515711] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:52.310 [2024-12-07 05:39:25.515729] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:52.310 [2024-12-07 05:39:25.515845] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:52.310 "name": "raid_bdev1", 00:11:52.310 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:11:52.310 "strip_size_kb": 0, 00:11:52.310 "state": "online", 00:11:52.310 "raid_level": "raid1", 00:11:52.310 "superblock": false, 00:11:52.310 "num_base_bdevs": 4, 00:11:52.310 "num_base_bdevs_discovered": 4, 00:11:52.310 "num_base_bdevs_operational": 4, 00:11:52.310 "base_bdevs_list": [ 00:11:52.310 { 00:11:52.310 "name": "BaseBdev1", 00:11:52.310 "uuid": "60518fd1-a66d-57e6-acc7-d4285d575b9f", 00:11:52.310 "is_configured": true, 00:11:52.310 "data_offset": 0, 00:11:52.310 "data_size": 65536 00:11:52.310 }, 00:11:52.310 { 00:11:52.310 "name": "BaseBdev2", 00:11:52.310 "uuid": "9b3b8679-e960-56b1-97d2-5783ece21c80", 00:11:52.310 "is_configured": true, 00:11:52.310 "data_offset": 0, 00:11:52.310 "data_size": 65536 00:11:52.310 }, 00:11:52.310 { 00:11:52.310 "name": "BaseBdev3", 00:11:52.310 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:11:52.310 "is_configured": true, 00:11:52.310 "data_offset": 0, 00:11:52.310 "data_size": 65536 00:11:52.310 }, 00:11:52.310 { 00:11:52.310 "name": "BaseBdev4", 00:11:52.310 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:11:52.310 "is_configured": true, 00:11:52.310 "data_offset": 0, 00:11:52.310 "data_size": 65536 00:11:52.310 } 00:11:52.310 ] 00:11:52.310 }' 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:52.310 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.879 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:52.879 05:39:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:52.879 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.879 05:39:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.879 [2024-12-07 05:39:25.980604] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:52.879 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.879 05:39:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:52.880 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:52.880 [2024-12-07 05:39:26.243891] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:53.140 /dev/nbd0 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:53.140 1+0 records in 00:11:53.140 1+0 records out 00:11:53.140 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000535378 s, 7.7 MB/s 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:53.140 05:39:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:11:58.436 65536+0 records in 00:11:58.436 65536+0 records out 00:11:58.436 33554432 bytes (34 MB, 32 MiB) copied, 5.38241 s, 6.2 MB/s 00:11:58.436 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:58.436 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:58.436 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:58.436 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:58.436 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:58.436 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:58.436 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:58.695 [2024-12-07 05:39:31.882366] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:58.695 [2024-12-07 05:39:31.918376] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.695 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:58.695 "name": "raid_bdev1", 00:11:58.695 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:11:58.695 "strip_size_kb": 0, 00:11:58.695 "state": "online", 00:11:58.696 "raid_level": "raid1", 00:11:58.696 "superblock": false, 00:11:58.696 "num_base_bdevs": 4, 00:11:58.696 "num_base_bdevs_discovered": 3, 00:11:58.696 "num_base_bdevs_operational": 3, 00:11:58.696 "base_bdevs_list": [ 00:11:58.696 { 00:11:58.696 "name": null, 00:11:58.696 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.696 "is_configured": false, 00:11:58.696 "data_offset": 0, 00:11:58.696 "data_size": 65536 00:11:58.696 }, 00:11:58.696 { 00:11:58.696 "name": "BaseBdev2", 00:11:58.696 "uuid": "9b3b8679-e960-56b1-97d2-5783ece21c80", 00:11:58.696 "is_configured": true, 00:11:58.696 "data_offset": 0, 00:11:58.696 "data_size": 65536 00:11:58.696 }, 00:11:58.696 { 00:11:58.696 "name": "BaseBdev3", 00:11:58.696 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:11:58.696 "is_configured": true, 00:11:58.696 "data_offset": 0, 00:11:58.696 "data_size": 65536 00:11:58.696 }, 00:11:58.696 { 00:11:58.696 "name": "BaseBdev4", 00:11:58.696 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:11:58.696 "is_configured": true, 00:11:58.696 "data_offset": 0, 00:11:58.696 "data_size": 65536 00:11:58.696 } 00:11:58.696 ] 00:11:58.696 }' 00:11:58.696 05:39:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:58.696 05:39:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:59.264 05:39:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:59.264 05:39:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.264 05:39:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:59.264 [2024-12-07 05:39:32.385687] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:59.264 [2024-12-07 05:39:32.390280] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d063c0 00:11:59.264 05:39:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.264 05:39:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:59.264 [2024-12-07 05:39:32.392662] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:00.202 "name": "raid_bdev1", 00:12:00.202 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:12:00.202 "strip_size_kb": 0, 00:12:00.202 "state": "online", 00:12:00.202 "raid_level": "raid1", 00:12:00.202 "superblock": false, 00:12:00.202 "num_base_bdevs": 4, 00:12:00.202 "num_base_bdevs_discovered": 4, 00:12:00.202 "num_base_bdevs_operational": 4, 00:12:00.202 "process": { 00:12:00.202 "type": "rebuild", 00:12:00.202 "target": "spare", 00:12:00.202 "progress": { 00:12:00.202 "blocks": 20480, 00:12:00.202 "percent": 31 00:12:00.202 } 00:12:00.202 }, 00:12:00.202 "base_bdevs_list": [ 00:12:00.202 { 00:12:00.202 "name": "spare", 00:12:00.202 "uuid": "7237df7d-a204-5b50-8277-d3988dfd2593", 00:12:00.202 "is_configured": true, 00:12:00.202 "data_offset": 0, 00:12:00.202 "data_size": 65536 00:12:00.202 }, 00:12:00.202 { 00:12:00.202 "name": "BaseBdev2", 00:12:00.202 "uuid": "9b3b8679-e960-56b1-97d2-5783ece21c80", 00:12:00.202 "is_configured": true, 00:12:00.202 "data_offset": 0, 00:12:00.202 "data_size": 65536 00:12:00.202 }, 00:12:00.202 { 00:12:00.202 "name": "BaseBdev3", 00:12:00.202 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:12:00.202 "is_configured": true, 00:12:00.202 "data_offset": 0, 00:12:00.202 "data_size": 65536 00:12:00.202 }, 00:12:00.202 { 00:12:00.202 "name": "BaseBdev4", 00:12:00.202 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:12:00.202 "is_configured": true, 00:12:00.202 "data_offset": 0, 00:12:00.202 "data_size": 65536 00:12:00.202 } 00:12:00.202 ] 00:12:00.202 }' 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.202 05:39:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:00.202 [2024-12-07 05:39:33.549728] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:00.461 [2024-12-07 05:39:33.598873] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:00.461 [2024-12-07 05:39:33.599001] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:00.461 [2024-12-07 05:39:33.599027] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:00.461 [2024-12-07 05:39:33.599037] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:00.461 05:39:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.461 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:00.461 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:00.461 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:00.461 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:00.462 "name": "raid_bdev1", 00:12:00.462 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:12:00.462 "strip_size_kb": 0, 00:12:00.462 "state": "online", 00:12:00.462 "raid_level": "raid1", 00:12:00.462 "superblock": false, 00:12:00.462 "num_base_bdevs": 4, 00:12:00.462 "num_base_bdevs_discovered": 3, 00:12:00.462 "num_base_bdevs_operational": 3, 00:12:00.462 "base_bdevs_list": [ 00:12:00.462 { 00:12:00.462 "name": null, 00:12:00.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:00.462 "is_configured": false, 00:12:00.462 "data_offset": 0, 00:12:00.462 "data_size": 65536 00:12:00.462 }, 00:12:00.462 { 00:12:00.462 "name": "BaseBdev2", 00:12:00.462 "uuid": "9b3b8679-e960-56b1-97d2-5783ece21c80", 00:12:00.462 "is_configured": true, 00:12:00.462 "data_offset": 0, 00:12:00.462 "data_size": 65536 00:12:00.462 }, 00:12:00.462 { 00:12:00.462 "name": "BaseBdev3", 00:12:00.462 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:12:00.462 "is_configured": true, 00:12:00.462 "data_offset": 0, 00:12:00.462 "data_size": 65536 00:12:00.462 }, 00:12:00.462 { 00:12:00.462 "name": "BaseBdev4", 00:12:00.462 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:12:00.462 "is_configured": true, 00:12:00.462 "data_offset": 0, 00:12:00.462 "data_size": 65536 00:12:00.462 } 00:12:00.462 ] 00:12:00.462 }' 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:00.462 05:39:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:01.029 "name": "raid_bdev1", 00:12:01.029 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:12:01.029 "strip_size_kb": 0, 00:12:01.029 "state": "online", 00:12:01.029 "raid_level": "raid1", 00:12:01.029 "superblock": false, 00:12:01.029 "num_base_bdevs": 4, 00:12:01.029 "num_base_bdevs_discovered": 3, 00:12:01.029 "num_base_bdevs_operational": 3, 00:12:01.029 "base_bdevs_list": [ 00:12:01.029 { 00:12:01.029 "name": null, 00:12:01.029 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:01.029 "is_configured": false, 00:12:01.029 "data_offset": 0, 00:12:01.029 "data_size": 65536 00:12:01.029 }, 00:12:01.029 { 00:12:01.029 "name": "BaseBdev2", 00:12:01.029 "uuid": "9b3b8679-e960-56b1-97d2-5783ece21c80", 00:12:01.029 "is_configured": true, 00:12:01.029 "data_offset": 0, 00:12:01.029 "data_size": 65536 00:12:01.029 }, 00:12:01.029 { 00:12:01.029 "name": "BaseBdev3", 00:12:01.029 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:12:01.029 "is_configured": true, 00:12:01.029 "data_offset": 0, 00:12:01.029 "data_size": 65536 00:12:01.029 }, 00:12:01.029 { 00:12:01.029 "name": "BaseBdev4", 00:12:01.029 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:12:01.029 "is_configured": true, 00:12:01.029 "data_offset": 0, 00:12:01.029 "data_size": 65536 00:12:01.029 } 00:12:01.029 ] 00:12:01.029 }' 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.029 [2024-12-07 05:39:34.275095] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:01.029 [2024-12-07 05:39:34.279747] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06490 00:12:01.029 05:39:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.030 05:39:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:01.030 [2024-12-07 05:39:34.282075] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:01.965 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:01.965 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:01.965 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:01.965 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:01.965 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:01.965 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.965 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:01.965 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.965 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:01.965 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:02.224 "name": "raid_bdev1", 00:12:02.224 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:12:02.224 "strip_size_kb": 0, 00:12:02.224 "state": "online", 00:12:02.224 "raid_level": "raid1", 00:12:02.224 "superblock": false, 00:12:02.224 "num_base_bdevs": 4, 00:12:02.224 "num_base_bdevs_discovered": 4, 00:12:02.224 "num_base_bdevs_operational": 4, 00:12:02.224 "process": { 00:12:02.224 "type": "rebuild", 00:12:02.224 "target": "spare", 00:12:02.224 "progress": { 00:12:02.224 "blocks": 20480, 00:12:02.224 "percent": 31 00:12:02.224 } 00:12:02.224 }, 00:12:02.224 "base_bdevs_list": [ 00:12:02.224 { 00:12:02.224 "name": "spare", 00:12:02.224 "uuid": "7237df7d-a204-5b50-8277-d3988dfd2593", 00:12:02.224 "is_configured": true, 00:12:02.224 "data_offset": 0, 00:12:02.224 "data_size": 65536 00:12:02.224 }, 00:12:02.224 { 00:12:02.224 "name": "BaseBdev2", 00:12:02.224 "uuid": "9b3b8679-e960-56b1-97d2-5783ece21c80", 00:12:02.224 "is_configured": true, 00:12:02.224 "data_offset": 0, 00:12:02.224 "data_size": 65536 00:12:02.224 }, 00:12:02.224 { 00:12:02.224 "name": "BaseBdev3", 00:12:02.224 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:12:02.224 "is_configured": true, 00:12:02.224 "data_offset": 0, 00:12:02.224 "data_size": 65536 00:12:02.224 }, 00:12:02.224 { 00:12:02.224 "name": "BaseBdev4", 00:12:02.224 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:12:02.224 "is_configured": true, 00:12:02.224 "data_offset": 0, 00:12:02.224 "data_size": 65536 00:12:02.224 } 00:12:02.224 ] 00:12:02.224 }' 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:02.224 [2024-12-07 05:39:35.450785] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:02.224 [2024-12-07 05:39:35.487780] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d06490 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:02.224 "name": "raid_bdev1", 00:12:02.224 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:12:02.224 "strip_size_kb": 0, 00:12:02.224 "state": "online", 00:12:02.224 "raid_level": "raid1", 00:12:02.224 "superblock": false, 00:12:02.224 "num_base_bdevs": 4, 00:12:02.224 "num_base_bdevs_discovered": 3, 00:12:02.224 "num_base_bdevs_operational": 3, 00:12:02.224 "process": { 00:12:02.224 "type": "rebuild", 00:12:02.224 "target": "spare", 00:12:02.224 "progress": { 00:12:02.224 "blocks": 24576, 00:12:02.224 "percent": 37 00:12:02.224 } 00:12:02.224 }, 00:12:02.224 "base_bdevs_list": [ 00:12:02.224 { 00:12:02.224 "name": "spare", 00:12:02.224 "uuid": "7237df7d-a204-5b50-8277-d3988dfd2593", 00:12:02.224 "is_configured": true, 00:12:02.224 "data_offset": 0, 00:12:02.224 "data_size": 65536 00:12:02.224 }, 00:12:02.224 { 00:12:02.224 "name": null, 00:12:02.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:02.224 "is_configured": false, 00:12:02.224 "data_offset": 0, 00:12:02.224 "data_size": 65536 00:12:02.224 }, 00:12:02.224 { 00:12:02.224 "name": "BaseBdev3", 00:12:02.224 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:12:02.224 "is_configured": true, 00:12:02.224 "data_offset": 0, 00:12:02.224 "data_size": 65536 00:12:02.224 }, 00:12:02.224 { 00:12:02.224 "name": "BaseBdev4", 00:12:02.224 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:12:02.224 "is_configured": true, 00:12:02.224 "data_offset": 0, 00:12:02.224 "data_size": 65536 00:12:02.224 } 00:12:02.224 ] 00:12:02.224 }' 00:12:02.224 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=352 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:02.483 "name": "raid_bdev1", 00:12:02.483 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:12:02.483 "strip_size_kb": 0, 00:12:02.483 "state": "online", 00:12:02.483 "raid_level": "raid1", 00:12:02.483 "superblock": false, 00:12:02.483 "num_base_bdevs": 4, 00:12:02.483 "num_base_bdevs_discovered": 3, 00:12:02.483 "num_base_bdevs_operational": 3, 00:12:02.483 "process": { 00:12:02.483 "type": "rebuild", 00:12:02.483 "target": "spare", 00:12:02.483 "progress": { 00:12:02.483 "blocks": 26624, 00:12:02.483 "percent": 40 00:12:02.483 } 00:12:02.483 }, 00:12:02.483 "base_bdevs_list": [ 00:12:02.483 { 00:12:02.483 "name": "spare", 00:12:02.483 "uuid": "7237df7d-a204-5b50-8277-d3988dfd2593", 00:12:02.483 "is_configured": true, 00:12:02.483 "data_offset": 0, 00:12:02.483 "data_size": 65536 00:12:02.483 }, 00:12:02.483 { 00:12:02.483 "name": null, 00:12:02.483 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:02.483 "is_configured": false, 00:12:02.483 "data_offset": 0, 00:12:02.483 "data_size": 65536 00:12:02.483 }, 00:12:02.483 { 00:12:02.483 "name": "BaseBdev3", 00:12:02.483 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:12:02.483 "is_configured": true, 00:12:02.483 "data_offset": 0, 00:12:02.483 "data_size": 65536 00:12:02.483 }, 00:12:02.483 { 00:12:02.483 "name": "BaseBdev4", 00:12:02.483 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:12:02.483 "is_configured": true, 00:12:02.483 "data_offset": 0, 00:12:02.483 "data_size": 65536 00:12:02.483 } 00:12:02.483 ] 00:12:02.483 }' 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:02.483 05:39:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:03.863 "name": "raid_bdev1", 00:12:03.863 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:12:03.863 "strip_size_kb": 0, 00:12:03.863 "state": "online", 00:12:03.863 "raid_level": "raid1", 00:12:03.863 "superblock": false, 00:12:03.863 "num_base_bdevs": 4, 00:12:03.863 "num_base_bdevs_discovered": 3, 00:12:03.863 "num_base_bdevs_operational": 3, 00:12:03.863 "process": { 00:12:03.863 "type": "rebuild", 00:12:03.863 "target": "spare", 00:12:03.863 "progress": { 00:12:03.863 "blocks": 51200, 00:12:03.863 "percent": 78 00:12:03.863 } 00:12:03.863 }, 00:12:03.863 "base_bdevs_list": [ 00:12:03.863 { 00:12:03.863 "name": "spare", 00:12:03.863 "uuid": "7237df7d-a204-5b50-8277-d3988dfd2593", 00:12:03.863 "is_configured": true, 00:12:03.863 "data_offset": 0, 00:12:03.863 "data_size": 65536 00:12:03.863 }, 00:12:03.863 { 00:12:03.863 "name": null, 00:12:03.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:03.863 "is_configured": false, 00:12:03.863 "data_offset": 0, 00:12:03.863 "data_size": 65536 00:12:03.863 }, 00:12:03.863 { 00:12:03.863 "name": "BaseBdev3", 00:12:03.863 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:12:03.863 "is_configured": true, 00:12:03.863 "data_offset": 0, 00:12:03.863 "data_size": 65536 00:12:03.863 }, 00:12:03.863 { 00:12:03.863 "name": "BaseBdev4", 00:12:03.863 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:12:03.863 "is_configured": true, 00:12:03.863 "data_offset": 0, 00:12:03.863 "data_size": 65536 00:12:03.863 } 00:12:03.863 ] 00:12:03.863 }' 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:03.863 05:39:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:04.432 [2024-12-07 05:39:37.497152] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:04.432 [2024-12-07 05:39:37.497270] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:04.432 [2024-12-07 05:39:37.497320] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:04.691 05:39:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:04.691 05:39:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:04.691 05:39:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:04.691 05:39:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:04.691 05:39:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:04.691 05:39:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:04.691 05:39:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.691 05:39:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:04.691 05:39:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.691 05:39:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:04.691 05:39:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.691 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:04.691 "name": "raid_bdev1", 00:12:04.691 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:12:04.691 "strip_size_kb": 0, 00:12:04.691 "state": "online", 00:12:04.691 "raid_level": "raid1", 00:12:04.691 "superblock": false, 00:12:04.691 "num_base_bdevs": 4, 00:12:04.691 "num_base_bdevs_discovered": 3, 00:12:04.691 "num_base_bdevs_operational": 3, 00:12:04.691 "base_bdevs_list": [ 00:12:04.691 { 00:12:04.691 "name": "spare", 00:12:04.691 "uuid": "7237df7d-a204-5b50-8277-d3988dfd2593", 00:12:04.691 "is_configured": true, 00:12:04.691 "data_offset": 0, 00:12:04.691 "data_size": 65536 00:12:04.691 }, 00:12:04.691 { 00:12:04.691 "name": null, 00:12:04.691 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:04.691 "is_configured": false, 00:12:04.691 "data_offset": 0, 00:12:04.691 "data_size": 65536 00:12:04.691 }, 00:12:04.691 { 00:12:04.691 "name": "BaseBdev3", 00:12:04.691 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:12:04.691 "is_configured": true, 00:12:04.691 "data_offset": 0, 00:12:04.691 "data_size": 65536 00:12:04.691 }, 00:12:04.691 { 00:12:04.691 "name": "BaseBdev4", 00:12:04.691 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:12:04.691 "is_configured": true, 00:12:04.691 "data_offset": 0, 00:12:04.691 "data_size": 65536 00:12:04.691 } 00:12:04.691 ] 00:12:04.691 }' 00:12:04.691 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:04.951 "name": "raid_bdev1", 00:12:04.951 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:12:04.951 "strip_size_kb": 0, 00:12:04.951 "state": "online", 00:12:04.951 "raid_level": "raid1", 00:12:04.951 "superblock": false, 00:12:04.951 "num_base_bdevs": 4, 00:12:04.951 "num_base_bdevs_discovered": 3, 00:12:04.951 "num_base_bdevs_operational": 3, 00:12:04.951 "base_bdevs_list": [ 00:12:04.951 { 00:12:04.951 "name": "spare", 00:12:04.951 "uuid": "7237df7d-a204-5b50-8277-d3988dfd2593", 00:12:04.951 "is_configured": true, 00:12:04.951 "data_offset": 0, 00:12:04.951 "data_size": 65536 00:12:04.951 }, 00:12:04.951 { 00:12:04.951 "name": null, 00:12:04.951 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:04.951 "is_configured": false, 00:12:04.951 "data_offset": 0, 00:12:04.951 "data_size": 65536 00:12:04.951 }, 00:12:04.951 { 00:12:04.951 "name": "BaseBdev3", 00:12:04.951 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:12:04.951 "is_configured": true, 00:12:04.951 "data_offset": 0, 00:12:04.951 "data_size": 65536 00:12:04.951 }, 00:12:04.951 { 00:12:04.951 "name": "BaseBdev4", 00:12:04.951 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:12:04.951 "is_configured": true, 00:12:04.951 "data_offset": 0, 00:12:04.951 "data_size": 65536 00:12:04.951 } 00:12:04.951 ] 00:12:04.951 }' 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:04.951 "name": "raid_bdev1", 00:12:04.951 "uuid": "dbb333b2-70a0-4c8e-840d-5a31164c1534", 00:12:04.951 "strip_size_kb": 0, 00:12:04.951 "state": "online", 00:12:04.951 "raid_level": "raid1", 00:12:04.951 "superblock": false, 00:12:04.951 "num_base_bdevs": 4, 00:12:04.951 "num_base_bdevs_discovered": 3, 00:12:04.951 "num_base_bdevs_operational": 3, 00:12:04.951 "base_bdevs_list": [ 00:12:04.951 { 00:12:04.951 "name": "spare", 00:12:04.951 "uuid": "7237df7d-a204-5b50-8277-d3988dfd2593", 00:12:04.951 "is_configured": true, 00:12:04.951 "data_offset": 0, 00:12:04.951 "data_size": 65536 00:12:04.951 }, 00:12:04.951 { 00:12:04.951 "name": null, 00:12:04.951 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:04.951 "is_configured": false, 00:12:04.951 "data_offset": 0, 00:12:04.951 "data_size": 65536 00:12:04.951 }, 00:12:04.951 { 00:12:04.951 "name": "BaseBdev3", 00:12:04.951 "uuid": "83bbfcbe-7e08-514d-97f3-a4320ee196b0", 00:12:04.951 "is_configured": true, 00:12:04.951 "data_offset": 0, 00:12:04.951 "data_size": 65536 00:12:04.951 }, 00:12:04.951 { 00:12:04.951 "name": "BaseBdev4", 00:12:04.951 "uuid": "74d1876b-a639-5f80-bada-b551c85311b1", 00:12:04.951 "is_configured": true, 00:12:04.951 "data_offset": 0, 00:12:04.951 "data_size": 65536 00:12:04.951 } 00:12:04.951 ] 00:12:04.951 }' 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:04.951 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:05.527 [2024-12-07 05:39:38.755803] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:05.527 [2024-12-07 05:39:38.755915] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:05.527 [2024-12-07 05:39:38.756027] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:05.527 [2024-12-07 05:39:38.756129] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:05.527 [2024-12-07 05:39:38.756150] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:05.527 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:05.528 05:39:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:05.528 05:39:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:05.528 05:39:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:05.528 05:39:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:05.528 05:39:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:05.528 05:39:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:05.528 05:39:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:05.528 05:39:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:05.528 05:39:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:05.528 05:39:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:05.807 /dev/nbd0 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:05.807 1+0 records in 00:12:05.807 1+0 records out 00:12:05.807 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000356096 s, 11.5 MB/s 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:05.807 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:06.079 /dev/nbd1 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:06.079 1+0 records in 00:12:06.079 1+0 records out 00:12:06.079 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000319062 s, 12.8 MB/s 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:06.079 05:39:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:06.337 05:39:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:06.337 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:06.337 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:06.337 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:06.337 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:06.337 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:06.337 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:06.596 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:06.596 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:06.596 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:06.596 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:06.596 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:06.596 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:06.596 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:06.596 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:06.596 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:06.596 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 87806 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 87806 ']' 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 87806 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:06.855 05:39:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87806 00:12:06.855 05:39:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:06.855 killing process with pid 87806 00:12:06.855 Received shutdown signal, test time was about 60.000000 seconds 00:12:06.855 00:12:06.855 Latency(us) 00:12:06.855 [2024-12-07T05:39:40.223Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:06.855 [2024-12-07T05:39:40.223Z] =================================================================================================================== 00:12:06.855 [2024-12-07T05:39:40.223Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:06.855 05:39:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:06.855 05:39:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87806' 00:12:06.855 05:39:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 87806 00:12:06.855 [2024-12-07 05:39:40.026292] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:06.855 05:39:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 87806 00:12:06.855 [2024-12-07 05:39:40.081041] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:12:07.113 ************************************ 00:12:07.113 END TEST raid_rebuild_test 00:12:07.113 ************************************ 00:12:07.113 00:12:07.113 real 0m15.866s 00:12:07.113 user 0m18.239s 00:12:07.113 sys 0m2.931s 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.113 05:39:40 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:12:07.113 05:39:40 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:07.113 05:39:40 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:07.113 05:39:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:07.113 ************************************ 00:12:07.113 START TEST raid_rebuild_test_sb 00:12:07.113 ************************************ 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true false true 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=88241 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 88241 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 88241 ']' 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:07.113 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:07.113 05:39:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.113 [2024-12-07 05:39:40.466843] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:12:07.114 [2024-12-07 05:39:40.467089] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88241 ] 00:12:07.114 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:07.114 Zero copy mechanism will not be used. 00:12:07.371 [2024-12-07 05:39:40.627051] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:07.371 [2024-12-07 05:39:40.657384] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:07.371 [2024-12-07 05:39:40.703462] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:07.371 [2024-12-07 05:39:40.703600] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 BaseBdev1_malloc 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 [2024-12-07 05:39:41.397921] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:08.306 [2024-12-07 05:39:41.397992] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.306 [2024-12-07 05:39:41.398028] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:08.306 [2024-12-07 05:39:41.398043] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.306 [2024-12-07 05:39:41.400584] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.306 [2024-12-07 05:39:41.400732] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:08.306 BaseBdev1 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 BaseBdev2_malloc 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 [2024-12-07 05:39:41.427385] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:08.306 [2024-12-07 05:39:41.427456] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.306 [2024-12-07 05:39:41.427487] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:08.306 [2024-12-07 05:39:41.427497] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.306 [2024-12-07 05:39:41.430052] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.306 [2024-12-07 05:39:41.430098] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:08.306 BaseBdev2 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 BaseBdev3_malloc 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 [2024-12-07 05:39:41.456947] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:08.306 [2024-12-07 05:39:41.457031] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.306 [2024-12-07 05:39:41.457058] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:08.306 [2024-12-07 05:39:41.457069] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.306 [2024-12-07 05:39:41.459647] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.306 [2024-12-07 05:39:41.459687] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:08.306 BaseBdev3 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 BaseBdev4_malloc 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 [2024-12-07 05:39:41.497726] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:08.306 [2024-12-07 05:39:41.497789] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.306 [2024-12-07 05:39:41.497837] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:08.306 [2024-12-07 05:39:41.497847] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.306 [2024-12-07 05:39:41.500387] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.306 [2024-12-07 05:39:41.500430] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:08.306 BaseBdev4 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 spare_malloc 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 spare_delay 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 [2024-12-07 05:39:41.539198] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:08.306 [2024-12-07 05:39:41.539321] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.306 [2024-12-07 05:39:41.539354] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:08.306 [2024-12-07 05:39:41.539366] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.306 [2024-12-07 05:39:41.541951] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.306 [2024-12-07 05:39:41.541993] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:08.306 spare 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 [2024-12-07 05:39:41.551270] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:08.306 [2024-12-07 05:39:41.553488] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:08.306 [2024-12-07 05:39:41.553561] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:08.306 [2024-12-07 05:39:41.553619] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:08.306 [2024-12-07 05:39:41.553843] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:08.306 [2024-12-07 05:39:41.553879] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:08.306 [2024-12-07 05:39:41.554178] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:08.306 [2024-12-07 05:39:41.554346] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:08.306 [2024-12-07 05:39:41.554366] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:08.306 [2024-12-07 05:39:41.554538] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.306 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:08.306 "name": "raid_bdev1", 00:12:08.306 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:08.306 "strip_size_kb": 0, 00:12:08.306 "state": "online", 00:12:08.306 "raid_level": "raid1", 00:12:08.306 "superblock": true, 00:12:08.306 "num_base_bdevs": 4, 00:12:08.306 "num_base_bdevs_discovered": 4, 00:12:08.306 "num_base_bdevs_operational": 4, 00:12:08.306 "base_bdevs_list": [ 00:12:08.306 { 00:12:08.307 "name": "BaseBdev1", 00:12:08.307 "uuid": "b081c394-81cd-551e-bbaa-b79823f36ac1", 00:12:08.307 "is_configured": true, 00:12:08.307 "data_offset": 2048, 00:12:08.307 "data_size": 63488 00:12:08.307 }, 00:12:08.307 { 00:12:08.307 "name": "BaseBdev2", 00:12:08.307 "uuid": "bd729153-c526-5225-9d88-66eee065b83d", 00:12:08.307 "is_configured": true, 00:12:08.307 "data_offset": 2048, 00:12:08.307 "data_size": 63488 00:12:08.307 }, 00:12:08.307 { 00:12:08.307 "name": "BaseBdev3", 00:12:08.307 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:08.307 "is_configured": true, 00:12:08.307 "data_offset": 2048, 00:12:08.307 "data_size": 63488 00:12:08.307 }, 00:12:08.307 { 00:12:08.307 "name": "BaseBdev4", 00:12:08.307 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:08.307 "is_configured": true, 00:12:08.307 "data_offset": 2048, 00:12:08.307 "data_size": 63488 00:12:08.307 } 00:12:08.307 ] 00:12:08.307 }' 00:12:08.307 05:39:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:08.307 05:39:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.874 [2024-12-07 05:39:42.047198] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:08.874 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:09.133 [2024-12-07 05:39:42.354496] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:12:09.133 /dev/nbd0 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:09.133 1+0 records in 00:12:09.133 1+0 records out 00:12:09.133 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000311836 s, 13.1 MB/s 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:09.133 05:39:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:12:14.416 63488+0 records in 00:12:14.416 63488+0 records out 00:12:14.416 32505856 bytes (33 MB, 31 MiB) copied, 5.13903 s, 6.3 MB/s 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:14.416 [2024-12-07 05:39:47.758040] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:14.416 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:14.676 [2024-12-07 05:39:47.790576] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:14.676 "name": "raid_bdev1", 00:12:14.676 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:14.676 "strip_size_kb": 0, 00:12:14.676 "state": "online", 00:12:14.676 "raid_level": "raid1", 00:12:14.676 "superblock": true, 00:12:14.676 "num_base_bdevs": 4, 00:12:14.676 "num_base_bdevs_discovered": 3, 00:12:14.676 "num_base_bdevs_operational": 3, 00:12:14.676 "base_bdevs_list": [ 00:12:14.676 { 00:12:14.676 "name": null, 00:12:14.676 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:14.676 "is_configured": false, 00:12:14.676 "data_offset": 0, 00:12:14.676 "data_size": 63488 00:12:14.676 }, 00:12:14.676 { 00:12:14.676 "name": "BaseBdev2", 00:12:14.676 "uuid": "bd729153-c526-5225-9d88-66eee065b83d", 00:12:14.676 "is_configured": true, 00:12:14.676 "data_offset": 2048, 00:12:14.676 "data_size": 63488 00:12:14.676 }, 00:12:14.676 { 00:12:14.676 "name": "BaseBdev3", 00:12:14.676 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:14.676 "is_configured": true, 00:12:14.676 "data_offset": 2048, 00:12:14.676 "data_size": 63488 00:12:14.676 }, 00:12:14.676 { 00:12:14.676 "name": "BaseBdev4", 00:12:14.676 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:14.676 "is_configured": true, 00:12:14.676 "data_offset": 2048, 00:12:14.676 "data_size": 63488 00:12:14.676 } 00:12:14.676 ] 00:12:14.676 }' 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:14.676 05:39:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:14.936 05:39:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:14.936 05:39:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.936 05:39:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:14.936 [2024-12-07 05:39:48.229805] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:14.936 [2024-12-07 05:39:48.233996] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e420 00:12:14.936 05:39:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.936 05:39:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:14.936 [2024-12-07 05:39:48.235937] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:16.318 "name": "raid_bdev1", 00:12:16.318 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:16.318 "strip_size_kb": 0, 00:12:16.318 "state": "online", 00:12:16.318 "raid_level": "raid1", 00:12:16.318 "superblock": true, 00:12:16.318 "num_base_bdevs": 4, 00:12:16.318 "num_base_bdevs_discovered": 4, 00:12:16.318 "num_base_bdevs_operational": 4, 00:12:16.318 "process": { 00:12:16.318 "type": "rebuild", 00:12:16.318 "target": "spare", 00:12:16.318 "progress": { 00:12:16.318 "blocks": 20480, 00:12:16.318 "percent": 32 00:12:16.318 } 00:12:16.318 }, 00:12:16.318 "base_bdevs_list": [ 00:12:16.318 { 00:12:16.318 "name": "spare", 00:12:16.318 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:16.318 "is_configured": true, 00:12:16.318 "data_offset": 2048, 00:12:16.318 "data_size": 63488 00:12:16.318 }, 00:12:16.318 { 00:12:16.318 "name": "BaseBdev2", 00:12:16.318 "uuid": "bd729153-c526-5225-9d88-66eee065b83d", 00:12:16.318 "is_configured": true, 00:12:16.318 "data_offset": 2048, 00:12:16.318 "data_size": 63488 00:12:16.318 }, 00:12:16.318 { 00:12:16.318 "name": "BaseBdev3", 00:12:16.318 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:16.318 "is_configured": true, 00:12:16.318 "data_offset": 2048, 00:12:16.318 "data_size": 63488 00:12:16.318 }, 00:12:16.318 { 00:12:16.318 "name": "BaseBdev4", 00:12:16.318 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:16.318 "is_configured": true, 00:12:16.318 "data_offset": 2048, 00:12:16.318 "data_size": 63488 00:12:16.318 } 00:12:16.318 ] 00:12:16.318 }' 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:16.318 [2024-12-07 05:39:49.397063] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:16.318 [2024-12-07 05:39:49.440640] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:16.318 [2024-12-07 05:39:49.440780] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:16.318 [2024-12-07 05:39:49.440826] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:16.318 [2024-12-07 05:39:49.440850] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.318 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:16.318 "name": "raid_bdev1", 00:12:16.318 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:16.318 "strip_size_kb": 0, 00:12:16.318 "state": "online", 00:12:16.318 "raid_level": "raid1", 00:12:16.318 "superblock": true, 00:12:16.318 "num_base_bdevs": 4, 00:12:16.318 "num_base_bdevs_discovered": 3, 00:12:16.318 "num_base_bdevs_operational": 3, 00:12:16.318 "base_bdevs_list": [ 00:12:16.318 { 00:12:16.318 "name": null, 00:12:16.318 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:16.318 "is_configured": false, 00:12:16.318 "data_offset": 0, 00:12:16.318 "data_size": 63488 00:12:16.318 }, 00:12:16.318 { 00:12:16.318 "name": "BaseBdev2", 00:12:16.318 "uuid": "bd729153-c526-5225-9d88-66eee065b83d", 00:12:16.318 "is_configured": true, 00:12:16.318 "data_offset": 2048, 00:12:16.318 "data_size": 63488 00:12:16.318 }, 00:12:16.318 { 00:12:16.318 "name": "BaseBdev3", 00:12:16.318 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:16.318 "is_configured": true, 00:12:16.318 "data_offset": 2048, 00:12:16.318 "data_size": 63488 00:12:16.318 }, 00:12:16.319 { 00:12:16.319 "name": "BaseBdev4", 00:12:16.319 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:16.319 "is_configured": true, 00:12:16.319 "data_offset": 2048, 00:12:16.319 "data_size": 63488 00:12:16.319 } 00:12:16.319 ] 00:12:16.319 }' 00:12:16.319 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:16.319 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:16.597 "name": "raid_bdev1", 00:12:16.597 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:16.597 "strip_size_kb": 0, 00:12:16.597 "state": "online", 00:12:16.597 "raid_level": "raid1", 00:12:16.597 "superblock": true, 00:12:16.597 "num_base_bdevs": 4, 00:12:16.597 "num_base_bdevs_discovered": 3, 00:12:16.597 "num_base_bdevs_operational": 3, 00:12:16.597 "base_bdevs_list": [ 00:12:16.597 { 00:12:16.597 "name": null, 00:12:16.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:16.597 "is_configured": false, 00:12:16.597 "data_offset": 0, 00:12:16.597 "data_size": 63488 00:12:16.597 }, 00:12:16.597 { 00:12:16.597 "name": "BaseBdev2", 00:12:16.597 "uuid": "bd729153-c526-5225-9d88-66eee065b83d", 00:12:16.597 "is_configured": true, 00:12:16.597 "data_offset": 2048, 00:12:16.597 "data_size": 63488 00:12:16.597 }, 00:12:16.597 { 00:12:16.597 "name": "BaseBdev3", 00:12:16.597 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:16.597 "is_configured": true, 00:12:16.597 "data_offset": 2048, 00:12:16.597 "data_size": 63488 00:12:16.597 }, 00:12:16.597 { 00:12:16.597 "name": "BaseBdev4", 00:12:16.597 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:16.597 "is_configured": true, 00:12:16.597 "data_offset": 2048, 00:12:16.597 "data_size": 63488 00:12:16.597 } 00:12:16.597 ] 00:12:16.597 }' 00:12:16.597 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:16.885 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:16.885 05:39:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:16.885 05:39:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:16.885 05:39:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:16.885 05:39:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.885 05:39:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:16.885 [2024-12-07 05:39:50.024558] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:16.885 [2024-12-07 05:39:50.028799] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e4f0 00:12:16.885 05:39:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.885 05:39:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:16.885 [2024-12-07 05:39:50.030762] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:17.825 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:17.825 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:17.825 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:17.825 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:17.825 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:17.825 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.825 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.825 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:17.825 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:17.825 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.825 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:17.825 "name": "raid_bdev1", 00:12:17.826 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:17.826 "strip_size_kb": 0, 00:12:17.826 "state": "online", 00:12:17.826 "raid_level": "raid1", 00:12:17.826 "superblock": true, 00:12:17.826 "num_base_bdevs": 4, 00:12:17.826 "num_base_bdevs_discovered": 4, 00:12:17.826 "num_base_bdevs_operational": 4, 00:12:17.826 "process": { 00:12:17.826 "type": "rebuild", 00:12:17.826 "target": "spare", 00:12:17.826 "progress": { 00:12:17.826 "blocks": 20480, 00:12:17.826 "percent": 32 00:12:17.826 } 00:12:17.826 }, 00:12:17.826 "base_bdevs_list": [ 00:12:17.826 { 00:12:17.826 "name": "spare", 00:12:17.826 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:17.826 "is_configured": true, 00:12:17.826 "data_offset": 2048, 00:12:17.826 "data_size": 63488 00:12:17.826 }, 00:12:17.826 { 00:12:17.826 "name": "BaseBdev2", 00:12:17.826 "uuid": "bd729153-c526-5225-9d88-66eee065b83d", 00:12:17.826 "is_configured": true, 00:12:17.826 "data_offset": 2048, 00:12:17.826 "data_size": 63488 00:12:17.826 }, 00:12:17.826 { 00:12:17.826 "name": "BaseBdev3", 00:12:17.826 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:17.826 "is_configured": true, 00:12:17.826 "data_offset": 2048, 00:12:17.826 "data_size": 63488 00:12:17.826 }, 00:12:17.826 { 00:12:17.826 "name": "BaseBdev4", 00:12:17.826 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:17.826 "is_configured": true, 00:12:17.826 "data_offset": 2048, 00:12:17.826 "data_size": 63488 00:12:17.826 } 00:12:17.826 ] 00:12:17.826 }' 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:17.826 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.826 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:18.086 [2024-12-07 05:39:51.191622] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:18.086 [2024-12-07 05:39:51.334685] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000c3e4f0 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:18.086 "name": "raid_bdev1", 00:12:18.086 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:18.086 "strip_size_kb": 0, 00:12:18.086 "state": "online", 00:12:18.086 "raid_level": "raid1", 00:12:18.086 "superblock": true, 00:12:18.086 "num_base_bdevs": 4, 00:12:18.086 "num_base_bdevs_discovered": 3, 00:12:18.086 "num_base_bdevs_operational": 3, 00:12:18.086 "process": { 00:12:18.086 "type": "rebuild", 00:12:18.086 "target": "spare", 00:12:18.086 "progress": { 00:12:18.086 "blocks": 24576, 00:12:18.086 "percent": 38 00:12:18.086 } 00:12:18.086 }, 00:12:18.086 "base_bdevs_list": [ 00:12:18.086 { 00:12:18.086 "name": "spare", 00:12:18.086 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:18.086 "is_configured": true, 00:12:18.086 "data_offset": 2048, 00:12:18.086 "data_size": 63488 00:12:18.086 }, 00:12:18.086 { 00:12:18.086 "name": null, 00:12:18.086 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:18.086 "is_configured": false, 00:12:18.086 "data_offset": 0, 00:12:18.086 "data_size": 63488 00:12:18.086 }, 00:12:18.086 { 00:12:18.086 "name": "BaseBdev3", 00:12:18.086 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:18.086 "is_configured": true, 00:12:18.086 "data_offset": 2048, 00:12:18.086 "data_size": 63488 00:12:18.086 }, 00:12:18.086 { 00:12:18.086 "name": "BaseBdev4", 00:12:18.086 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:18.086 "is_configured": true, 00:12:18.086 "data_offset": 2048, 00:12:18.086 "data_size": 63488 00:12:18.086 } 00:12:18.086 ] 00:12:18.086 }' 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:18.086 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=368 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:18.346 "name": "raid_bdev1", 00:12:18.346 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:18.346 "strip_size_kb": 0, 00:12:18.346 "state": "online", 00:12:18.346 "raid_level": "raid1", 00:12:18.346 "superblock": true, 00:12:18.346 "num_base_bdevs": 4, 00:12:18.346 "num_base_bdevs_discovered": 3, 00:12:18.346 "num_base_bdevs_operational": 3, 00:12:18.346 "process": { 00:12:18.346 "type": "rebuild", 00:12:18.346 "target": "spare", 00:12:18.346 "progress": { 00:12:18.346 "blocks": 26624, 00:12:18.346 "percent": 41 00:12:18.346 } 00:12:18.346 }, 00:12:18.346 "base_bdevs_list": [ 00:12:18.346 { 00:12:18.346 "name": "spare", 00:12:18.346 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:18.346 "is_configured": true, 00:12:18.346 "data_offset": 2048, 00:12:18.346 "data_size": 63488 00:12:18.346 }, 00:12:18.346 { 00:12:18.346 "name": null, 00:12:18.346 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:18.346 "is_configured": false, 00:12:18.346 "data_offset": 0, 00:12:18.346 "data_size": 63488 00:12:18.346 }, 00:12:18.346 { 00:12:18.346 "name": "BaseBdev3", 00:12:18.346 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:18.346 "is_configured": true, 00:12:18.346 "data_offset": 2048, 00:12:18.346 "data_size": 63488 00:12:18.346 }, 00:12:18.346 { 00:12:18.346 "name": "BaseBdev4", 00:12:18.346 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:18.346 "is_configured": true, 00:12:18.346 "data_offset": 2048, 00:12:18.346 "data_size": 63488 00:12:18.346 } 00:12:18.346 ] 00:12:18.346 }' 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:18.346 05:39:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:19.286 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:19.286 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:19.286 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:19.286 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:19.286 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:19.286 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:19.286 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:19.286 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:19.286 05:39:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.286 05:39:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:19.286 05:39:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.546 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:19.546 "name": "raid_bdev1", 00:12:19.546 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:19.546 "strip_size_kb": 0, 00:12:19.546 "state": "online", 00:12:19.546 "raid_level": "raid1", 00:12:19.546 "superblock": true, 00:12:19.546 "num_base_bdevs": 4, 00:12:19.546 "num_base_bdevs_discovered": 3, 00:12:19.546 "num_base_bdevs_operational": 3, 00:12:19.546 "process": { 00:12:19.546 "type": "rebuild", 00:12:19.546 "target": "spare", 00:12:19.546 "progress": { 00:12:19.546 "blocks": 49152, 00:12:19.546 "percent": 77 00:12:19.546 } 00:12:19.546 }, 00:12:19.546 "base_bdevs_list": [ 00:12:19.546 { 00:12:19.546 "name": "spare", 00:12:19.546 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:19.546 "is_configured": true, 00:12:19.546 "data_offset": 2048, 00:12:19.546 "data_size": 63488 00:12:19.546 }, 00:12:19.546 { 00:12:19.546 "name": null, 00:12:19.546 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:19.546 "is_configured": false, 00:12:19.546 "data_offset": 0, 00:12:19.546 "data_size": 63488 00:12:19.546 }, 00:12:19.546 { 00:12:19.546 "name": "BaseBdev3", 00:12:19.546 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:19.546 "is_configured": true, 00:12:19.546 "data_offset": 2048, 00:12:19.546 "data_size": 63488 00:12:19.546 }, 00:12:19.546 { 00:12:19.546 "name": "BaseBdev4", 00:12:19.546 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:19.546 "is_configured": true, 00:12:19.546 "data_offset": 2048, 00:12:19.546 "data_size": 63488 00:12:19.546 } 00:12:19.546 ] 00:12:19.546 }' 00:12:19.546 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:19.546 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:19.546 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:19.546 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:19.546 05:39:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:20.116 [2024-12-07 05:39:53.241708] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:20.116 [2024-12-07 05:39:53.241802] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:20.116 [2024-12-07 05:39:53.241924] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:20.684 "name": "raid_bdev1", 00:12:20.684 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:20.684 "strip_size_kb": 0, 00:12:20.684 "state": "online", 00:12:20.684 "raid_level": "raid1", 00:12:20.684 "superblock": true, 00:12:20.684 "num_base_bdevs": 4, 00:12:20.684 "num_base_bdevs_discovered": 3, 00:12:20.684 "num_base_bdevs_operational": 3, 00:12:20.684 "base_bdevs_list": [ 00:12:20.684 { 00:12:20.684 "name": "spare", 00:12:20.684 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:20.684 "is_configured": true, 00:12:20.684 "data_offset": 2048, 00:12:20.684 "data_size": 63488 00:12:20.684 }, 00:12:20.684 { 00:12:20.684 "name": null, 00:12:20.684 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:20.684 "is_configured": false, 00:12:20.684 "data_offset": 0, 00:12:20.684 "data_size": 63488 00:12:20.684 }, 00:12:20.684 { 00:12:20.684 "name": "BaseBdev3", 00:12:20.684 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:20.684 "is_configured": true, 00:12:20.684 "data_offset": 2048, 00:12:20.684 "data_size": 63488 00:12:20.684 }, 00:12:20.684 { 00:12:20.684 "name": "BaseBdev4", 00:12:20.684 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:20.684 "is_configured": true, 00:12:20.684 "data_offset": 2048, 00:12:20.684 "data_size": 63488 00:12:20.684 } 00:12:20.684 ] 00:12:20.684 }' 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:20.684 "name": "raid_bdev1", 00:12:20.684 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:20.684 "strip_size_kb": 0, 00:12:20.684 "state": "online", 00:12:20.684 "raid_level": "raid1", 00:12:20.684 "superblock": true, 00:12:20.684 "num_base_bdevs": 4, 00:12:20.684 "num_base_bdevs_discovered": 3, 00:12:20.684 "num_base_bdevs_operational": 3, 00:12:20.684 "base_bdevs_list": [ 00:12:20.684 { 00:12:20.684 "name": "spare", 00:12:20.684 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:20.684 "is_configured": true, 00:12:20.684 "data_offset": 2048, 00:12:20.684 "data_size": 63488 00:12:20.684 }, 00:12:20.684 { 00:12:20.684 "name": null, 00:12:20.684 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:20.684 "is_configured": false, 00:12:20.684 "data_offset": 0, 00:12:20.684 "data_size": 63488 00:12:20.684 }, 00:12:20.684 { 00:12:20.684 "name": "BaseBdev3", 00:12:20.684 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:20.684 "is_configured": true, 00:12:20.684 "data_offset": 2048, 00:12:20.684 "data_size": 63488 00:12:20.684 }, 00:12:20.684 { 00:12:20.684 "name": "BaseBdev4", 00:12:20.684 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:20.684 "is_configured": true, 00:12:20.684 "data_offset": 2048, 00:12:20.684 "data_size": 63488 00:12:20.684 } 00:12:20.684 ] 00:12:20.684 }' 00:12:20.684 05:39:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:20.684 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:20.684 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:20.945 "name": "raid_bdev1", 00:12:20.945 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:20.945 "strip_size_kb": 0, 00:12:20.945 "state": "online", 00:12:20.945 "raid_level": "raid1", 00:12:20.945 "superblock": true, 00:12:20.945 "num_base_bdevs": 4, 00:12:20.945 "num_base_bdevs_discovered": 3, 00:12:20.945 "num_base_bdevs_operational": 3, 00:12:20.945 "base_bdevs_list": [ 00:12:20.945 { 00:12:20.945 "name": "spare", 00:12:20.945 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:20.945 "is_configured": true, 00:12:20.945 "data_offset": 2048, 00:12:20.945 "data_size": 63488 00:12:20.945 }, 00:12:20.945 { 00:12:20.945 "name": null, 00:12:20.945 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:20.945 "is_configured": false, 00:12:20.945 "data_offset": 0, 00:12:20.945 "data_size": 63488 00:12:20.945 }, 00:12:20.945 { 00:12:20.945 "name": "BaseBdev3", 00:12:20.945 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:20.945 "is_configured": true, 00:12:20.945 "data_offset": 2048, 00:12:20.945 "data_size": 63488 00:12:20.945 }, 00:12:20.945 { 00:12:20.945 "name": "BaseBdev4", 00:12:20.945 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:20.945 "is_configured": true, 00:12:20.945 "data_offset": 2048, 00:12:20.945 "data_size": 63488 00:12:20.945 } 00:12:20.945 ] 00:12:20.945 }' 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:20.945 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:21.205 [2024-12-07 05:39:54.535940] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:21.205 [2024-12-07 05:39:54.535971] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:21.205 [2024-12-07 05:39:54.536064] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:21.205 [2024-12-07 05:39:54.536151] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:21.205 [2024-12-07 05:39:54.536164] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:21.205 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:21.465 /dev/nbd0 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:21.465 1+0 records in 00:12:21.465 1+0 records out 00:12:21.465 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000225912 s, 18.1 MB/s 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:21.465 05:39:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:21.724 /dev/nbd1 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:21.724 1+0 records in 00:12:21.724 1+0 records out 00:12:21.724 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000228657 s, 17.9 MB/s 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:21.724 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:21.982 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:22.241 [2024-12-07 05:39:55.573494] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:22.241 [2024-12-07 05:39:55.573559] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:22.241 [2024-12-07 05:39:55.573581] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:12:22.241 [2024-12-07 05:39:55.573593] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:22.241 [2024-12-07 05:39:55.575813] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:22.241 [2024-12-07 05:39:55.575894] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:22.241 [2024-12-07 05:39:55.575985] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:22.241 [2024-12-07 05:39:55.576044] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:22.241 [2024-12-07 05:39:55.576152] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:22.241 [2024-12-07 05:39:55.576238] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:22.241 spare 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.241 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:22.501 [2024-12-07 05:39:55.676134] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:12:22.501 [2024-12-07 05:39:55.676170] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:22.501 [2024-12-07 05:39:55.676437] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:12:22.501 [2024-12-07 05:39:55.676577] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:12:22.501 [2024-12-07 05:39:55.676586] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:12:22.501 [2024-12-07 05:39:55.676794] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:22.501 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.501 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:22.501 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:22.501 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:22.501 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:22.501 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:22.501 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:22.502 "name": "raid_bdev1", 00:12:22.502 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:22.502 "strip_size_kb": 0, 00:12:22.502 "state": "online", 00:12:22.502 "raid_level": "raid1", 00:12:22.502 "superblock": true, 00:12:22.502 "num_base_bdevs": 4, 00:12:22.502 "num_base_bdevs_discovered": 3, 00:12:22.502 "num_base_bdevs_operational": 3, 00:12:22.502 "base_bdevs_list": [ 00:12:22.502 { 00:12:22.502 "name": "spare", 00:12:22.502 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:22.502 "is_configured": true, 00:12:22.502 "data_offset": 2048, 00:12:22.502 "data_size": 63488 00:12:22.502 }, 00:12:22.502 { 00:12:22.502 "name": null, 00:12:22.502 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:22.502 "is_configured": false, 00:12:22.502 "data_offset": 2048, 00:12:22.502 "data_size": 63488 00:12:22.502 }, 00:12:22.502 { 00:12:22.502 "name": "BaseBdev3", 00:12:22.502 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:22.502 "is_configured": true, 00:12:22.502 "data_offset": 2048, 00:12:22.502 "data_size": 63488 00:12:22.502 }, 00:12:22.502 { 00:12:22.502 "name": "BaseBdev4", 00:12:22.502 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:22.502 "is_configured": true, 00:12:22.502 "data_offset": 2048, 00:12:22.502 "data_size": 63488 00:12:22.502 } 00:12:22.502 ] 00:12:22.502 }' 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:22.502 05:39:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:23.072 "name": "raid_bdev1", 00:12:23.072 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:23.072 "strip_size_kb": 0, 00:12:23.072 "state": "online", 00:12:23.072 "raid_level": "raid1", 00:12:23.072 "superblock": true, 00:12:23.072 "num_base_bdevs": 4, 00:12:23.072 "num_base_bdevs_discovered": 3, 00:12:23.072 "num_base_bdevs_operational": 3, 00:12:23.072 "base_bdevs_list": [ 00:12:23.072 { 00:12:23.072 "name": "spare", 00:12:23.072 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:23.072 "is_configured": true, 00:12:23.072 "data_offset": 2048, 00:12:23.072 "data_size": 63488 00:12:23.072 }, 00:12:23.072 { 00:12:23.072 "name": null, 00:12:23.072 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:23.072 "is_configured": false, 00:12:23.072 "data_offset": 2048, 00:12:23.072 "data_size": 63488 00:12:23.072 }, 00:12:23.072 { 00:12:23.072 "name": "BaseBdev3", 00:12:23.072 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:23.072 "is_configured": true, 00:12:23.072 "data_offset": 2048, 00:12:23.072 "data_size": 63488 00:12:23.072 }, 00:12:23.072 { 00:12:23.072 "name": "BaseBdev4", 00:12:23.072 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:23.072 "is_configured": true, 00:12:23.072 "data_offset": 2048, 00:12:23.072 "data_size": 63488 00:12:23.072 } 00:12:23.072 ] 00:12:23.072 }' 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:23.072 [2024-12-07 05:39:56.320238] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.072 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:23.072 "name": "raid_bdev1", 00:12:23.072 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:23.072 "strip_size_kb": 0, 00:12:23.072 "state": "online", 00:12:23.072 "raid_level": "raid1", 00:12:23.072 "superblock": true, 00:12:23.072 "num_base_bdevs": 4, 00:12:23.072 "num_base_bdevs_discovered": 2, 00:12:23.072 "num_base_bdevs_operational": 2, 00:12:23.072 "base_bdevs_list": [ 00:12:23.072 { 00:12:23.072 "name": null, 00:12:23.073 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:23.073 "is_configured": false, 00:12:23.073 "data_offset": 0, 00:12:23.073 "data_size": 63488 00:12:23.073 }, 00:12:23.073 { 00:12:23.073 "name": null, 00:12:23.073 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:23.073 "is_configured": false, 00:12:23.073 "data_offset": 2048, 00:12:23.073 "data_size": 63488 00:12:23.073 }, 00:12:23.073 { 00:12:23.073 "name": "BaseBdev3", 00:12:23.073 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:23.073 "is_configured": true, 00:12:23.073 "data_offset": 2048, 00:12:23.073 "data_size": 63488 00:12:23.073 }, 00:12:23.073 { 00:12:23.073 "name": "BaseBdev4", 00:12:23.073 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:23.073 "is_configured": true, 00:12:23.073 "data_offset": 2048, 00:12:23.073 "data_size": 63488 00:12:23.073 } 00:12:23.073 ] 00:12:23.073 }' 00:12:23.073 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:23.073 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:23.333 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:23.333 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.333 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:23.333 [2024-12-07 05:39:56.655717] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:23.333 [2024-12-07 05:39:56.655956] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:23.333 [2024-12-07 05:39:56.656024] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:23.333 [2024-12-07 05:39:56.656109] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:23.333 [2024-12-07 05:39:56.660234] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caebd0 00:12:23.333 05:39:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.333 05:39:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:23.333 [2024-12-07 05:39:56.662225] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:24.714 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:24.714 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:24.714 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:24.714 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:24.714 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:24.714 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.714 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.714 05:39:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:24.714 05:39:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:24.714 05:39:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:24.714 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:24.714 "name": "raid_bdev1", 00:12:24.714 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:24.714 "strip_size_kb": 0, 00:12:24.714 "state": "online", 00:12:24.714 "raid_level": "raid1", 00:12:24.714 "superblock": true, 00:12:24.714 "num_base_bdevs": 4, 00:12:24.714 "num_base_bdevs_discovered": 3, 00:12:24.714 "num_base_bdevs_operational": 3, 00:12:24.714 "process": { 00:12:24.714 "type": "rebuild", 00:12:24.714 "target": "spare", 00:12:24.714 "progress": { 00:12:24.714 "blocks": 20480, 00:12:24.714 "percent": 32 00:12:24.714 } 00:12:24.714 }, 00:12:24.714 "base_bdevs_list": [ 00:12:24.714 { 00:12:24.714 "name": "spare", 00:12:24.714 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:24.714 "is_configured": true, 00:12:24.714 "data_offset": 2048, 00:12:24.714 "data_size": 63488 00:12:24.714 }, 00:12:24.714 { 00:12:24.715 "name": null, 00:12:24.715 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:24.715 "is_configured": false, 00:12:24.715 "data_offset": 2048, 00:12:24.715 "data_size": 63488 00:12:24.715 }, 00:12:24.715 { 00:12:24.715 "name": "BaseBdev3", 00:12:24.715 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:24.715 "is_configured": true, 00:12:24.715 "data_offset": 2048, 00:12:24.715 "data_size": 63488 00:12:24.715 }, 00:12:24.715 { 00:12:24.715 "name": "BaseBdev4", 00:12:24.715 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:24.715 "is_configured": true, 00:12:24.715 "data_offset": 2048, 00:12:24.715 "data_size": 63488 00:12:24.715 } 00:12:24.715 ] 00:12:24.715 }' 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:24.715 [2024-12-07 05:39:57.802776] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:24.715 [2024-12-07 05:39:57.866255] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:24.715 [2024-12-07 05:39:57.866317] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:24.715 [2024-12-07 05:39:57.866333] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:24.715 [2024-12-07 05:39:57.866342] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:24.715 "name": "raid_bdev1", 00:12:24.715 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:24.715 "strip_size_kb": 0, 00:12:24.715 "state": "online", 00:12:24.715 "raid_level": "raid1", 00:12:24.715 "superblock": true, 00:12:24.715 "num_base_bdevs": 4, 00:12:24.715 "num_base_bdevs_discovered": 2, 00:12:24.715 "num_base_bdevs_operational": 2, 00:12:24.715 "base_bdevs_list": [ 00:12:24.715 { 00:12:24.715 "name": null, 00:12:24.715 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:24.715 "is_configured": false, 00:12:24.715 "data_offset": 0, 00:12:24.715 "data_size": 63488 00:12:24.715 }, 00:12:24.715 { 00:12:24.715 "name": null, 00:12:24.715 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:24.715 "is_configured": false, 00:12:24.715 "data_offset": 2048, 00:12:24.715 "data_size": 63488 00:12:24.715 }, 00:12:24.715 { 00:12:24.715 "name": "BaseBdev3", 00:12:24.715 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:24.715 "is_configured": true, 00:12:24.715 "data_offset": 2048, 00:12:24.715 "data_size": 63488 00:12:24.715 }, 00:12:24.715 { 00:12:24.715 "name": "BaseBdev4", 00:12:24.715 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:24.715 "is_configured": true, 00:12:24.715 "data_offset": 2048, 00:12:24.715 "data_size": 63488 00:12:24.715 } 00:12:24.715 ] 00:12:24.715 }' 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:24.715 05:39:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:24.975 05:39:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:24.975 05:39:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:24.975 05:39:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:24.975 [2024-12-07 05:39:58.305852] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:24.975 [2024-12-07 05:39:58.305961] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:24.975 [2024-12-07 05:39:58.306005] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:12:24.975 [2024-12-07 05:39:58.306069] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:24.975 [2024-12-07 05:39:58.306554] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:24.975 [2024-12-07 05:39:58.306618] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:24.975 [2024-12-07 05:39:58.306752] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:24.975 [2024-12-07 05:39:58.306797] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:24.975 [2024-12-07 05:39:58.306838] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:24.975 [2024-12-07 05:39:58.306911] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:24.975 [2024-12-07 05:39:58.310926] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeca0 00:12:24.975 spare 00:12:24.975 05:39:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:24.975 05:39:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:24.975 [2024-12-07 05:39:58.312895] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:26.355 "name": "raid_bdev1", 00:12:26.355 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:26.355 "strip_size_kb": 0, 00:12:26.355 "state": "online", 00:12:26.355 "raid_level": "raid1", 00:12:26.355 "superblock": true, 00:12:26.355 "num_base_bdevs": 4, 00:12:26.355 "num_base_bdevs_discovered": 3, 00:12:26.355 "num_base_bdevs_operational": 3, 00:12:26.355 "process": { 00:12:26.355 "type": "rebuild", 00:12:26.355 "target": "spare", 00:12:26.355 "progress": { 00:12:26.355 "blocks": 20480, 00:12:26.355 "percent": 32 00:12:26.355 } 00:12:26.355 }, 00:12:26.355 "base_bdevs_list": [ 00:12:26.355 { 00:12:26.355 "name": "spare", 00:12:26.355 "uuid": "09ca6f17-452d-53e0-a327-fc8d2230c821", 00:12:26.355 "is_configured": true, 00:12:26.355 "data_offset": 2048, 00:12:26.355 "data_size": 63488 00:12:26.355 }, 00:12:26.355 { 00:12:26.355 "name": null, 00:12:26.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:26.355 "is_configured": false, 00:12:26.355 "data_offset": 2048, 00:12:26.355 "data_size": 63488 00:12:26.355 }, 00:12:26.355 { 00:12:26.355 "name": "BaseBdev3", 00:12:26.355 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:26.355 "is_configured": true, 00:12:26.355 "data_offset": 2048, 00:12:26.355 "data_size": 63488 00:12:26.355 }, 00:12:26.355 { 00:12:26.355 "name": "BaseBdev4", 00:12:26.355 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:26.355 "is_configured": true, 00:12:26.355 "data_offset": 2048, 00:12:26.355 "data_size": 63488 00:12:26.355 } 00:12:26.355 ] 00:12:26.355 }' 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.355 [2024-12-07 05:39:59.453179] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:26.355 [2024-12-07 05:39:59.516992] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:26.355 [2024-12-07 05:39:59.517044] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:26.355 [2024-12-07 05:39:59.517062] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:26.355 [2024-12-07 05:39:59.517069] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:26.355 "name": "raid_bdev1", 00:12:26.355 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:26.355 "strip_size_kb": 0, 00:12:26.355 "state": "online", 00:12:26.355 "raid_level": "raid1", 00:12:26.355 "superblock": true, 00:12:26.355 "num_base_bdevs": 4, 00:12:26.355 "num_base_bdevs_discovered": 2, 00:12:26.355 "num_base_bdevs_operational": 2, 00:12:26.355 "base_bdevs_list": [ 00:12:26.355 { 00:12:26.355 "name": null, 00:12:26.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:26.355 "is_configured": false, 00:12:26.355 "data_offset": 0, 00:12:26.355 "data_size": 63488 00:12:26.355 }, 00:12:26.355 { 00:12:26.355 "name": null, 00:12:26.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:26.355 "is_configured": false, 00:12:26.355 "data_offset": 2048, 00:12:26.355 "data_size": 63488 00:12:26.355 }, 00:12:26.355 { 00:12:26.355 "name": "BaseBdev3", 00:12:26.355 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:26.355 "is_configured": true, 00:12:26.355 "data_offset": 2048, 00:12:26.355 "data_size": 63488 00:12:26.355 }, 00:12:26.355 { 00:12:26.355 "name": "BaseBdev4", 00:12:26.355 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:26.355 "is_configured": true, 00:12:26.355 "data_offset": 2048, 00:12:26.355 "data_size": 63488 00:12:26.355 } 00:12:26.355 ] 00:12:26.355 }' 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:26.355 05:39:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.923 05:39:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:26.923 "name": "raid_bdev1", 00:12:26.923 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:26.923 "strip_size_kb": 0, 00:12:26.923 "state": "online", 00:12:26.923 "raid_level": "raid1", 00:12:26.923 "superblock": true, 00:12:26.923 "num_base_bdevs": 4, 00:12:26.923 "num_base_bdevs_discovered": 2, 00:12:26.923 "num_base_bdevs_operational": 2, 00:12:26.923 "base_bdevs_list": [ 00:12:26.923 { 00:12:26.923 "name": null, 00:12:26.923 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:26.923 "is_configured": false, 00:12:26.923 "data_offset": 0, 00:12:26.923 "data_size": 63488 00:12:26.923 }, 00:12:26.923 { 00:12:26.923 "name": null, 00:12:26.923 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:26.923 "is_configured": false, 00:12:26.923 "data_offset": 2048, 00:12:26.923 "data_size": 63488 00:12:26.923 }, 00:12:26.923 { 00:12:26.923 "name": "BaseBdev3", 00:12:26.923 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:26.923 "is_configured": true, 00:12:26.923 "data_offset": 2048, 00:12:26.923 "data_size": 63488 00:12:26.923 }, 00:12:26.923 { 00:12:26.923 "name": "BaseBdev4", 00:12:26.923 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:26.923 "is_configured": true, 00:12:26.923 "data_offset": 2048, 00:12:26.923 "data_size": 63488 00:12:26.923 } 00:12:26.923 ] 00:12:26.923 }' 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.923 [2024-12-07 05:40:00.104273] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:26.923 [2024-12-07 05:40:00.104330] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:26.923 [2024-12-07 05:40:00.104367] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:12:26.923 [2024-12-07 05:40:00.104375] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:26.923 [2024-12-07 05:40:00.104790] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:26.923 [2024-12-07 05:40:00.104808] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:26.923 [2024-12-07 05:40:00.104889] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:26.923 [2024-12-07 05:40:00.104903] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:26.923 [2024-12-07 05:40:00.104915] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:26.923 [2024-12-07 05:40:00.104924] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:26.923 BaseBdev1 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:26.923 05:40:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.914 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:27.914 "name": "raid_bdev1", 00:12:27.914 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:27.914 "strip_size_kb": 0, 00:12:27.914 "state": "online", 00:12:27.914 "raid_level": "raid1", 00:12:27.914 "superblock": true, 00:12:27.914 "num_base_bdevs": 4, 00:12:27.914 "num_base_bdevs_discovered": 2, 00:12:27.914 "num_base_bdevs_operational": 2, 00:12:27.914 "base_bdevs_list": [ 00:12:27.914 { 00:12:27.914 "name": null, 00:12:27.914 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:27.914 "is_configured": false, 00:12:27.914 "data_offset": 0, 00:12:27.914 "data_size": 63488 00:12:27.914 }, 00:12:27.914 { 00:12:27.914 "name": null, 00:12:27.914 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:27.914 "is_configured": false, 00:12:27.914 "data_offset": 2048, 00:12:27.914 "data_size": 63488 00:12:27.914 }, 00:12:27.914 { 00:12:27.914 "name": "BaseBdev3", 00:12:27.914 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:27.914 "is_configured": true, 00:12:27.914 "data_offset": 2048, 00:12:27.914 "data_size": 63488 00:12:27.914 }, 00:12:27.914 { 00:12:27.914 "name": "BaseBdev4", 00:12:27.914 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:27.914 "is_configured": true, 00:12:27.914 "data_offset": 2048, 00:12:27.914 "data_size": 63488 00:12:27.914 } 00:12:27.915 ] 00:12:27.915 }' 00:12:27.915 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:27.915 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:28.174 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:28.174 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:28.174 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:28.174 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:28.174 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:28.174 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:28.174 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:28.174 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:28.174 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:28.174 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:28.433 "name": "raid_bdev1", 00:12:28.433 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:28.433 "strip_size_kb": 0, 00:12:28.433 "state": "online", 00:12:28.433 "raid_level": "raid1", 00:12:28.433 "superblock": true, 00:12:28.433 "num_base_bdevs": 4, 00:12:28.433 "num_base_bdevs_discovered": 2, 00:12:28.433 "num_base_bdevs_operational": 2, 00:12:28.433 "base_bdevs_list": [ 00:12:28.433 { 00:12:28.433 "name": null, 00:12:28.433 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:28.433 "is_configured": false, 00:12:28.433 "data_offset": 0, 00:12:28.433 "data_size": 63488 00:12:28.433 }, 00:12:28.433 { 00:12:28.433 "name": null, 00:12:28.433 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:28.433 "is_configured": false, 00:12:28.433 "data_offset": 2048, 00:12:28.433 "data_size": 63488 00:12:28.433 }, 00:12:28.433 { 00:12:28.433 "name": "BaseBdev3", 00:12:28.433 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:28.433 "is_configured": true, 00:12:28.433 "data_offset": 2048, 00:12:28.433 "data_size": 63488 00:12:28.433 }, 00:12:28.433 { 00:12:28.433 "name": "BaseBdev4", 00:12:28.433 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:28.433 "is_configured": true, 00:12:28.433 "data_offset": 2048, 00:12:28.433 "data_size": 63488 00:12:28.433 } 00:12:28.433 ] 00:12:28.433 }' 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:28.433 [2024-12-07 05:40:01.661984] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:28.433 [2024-12-07 05:40:01.662225] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:28.433 [2024-12-07 05:40:01.662256] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:28.433 request: 00:12:28.433 { 00:12:28.433 "base_bdev": "BaseBdev1", 00:12:28.433 "raid_bdev": "raid_bdev1", 00:12:28.433 "method": "bdev_raid_add_base_bdev", 00:12:28.433 "req_id": 1 00:12:28.433 } 00:12:28.433 Got JSON-RPC error response 00:12:28.433 response: 00:12:28.433 { 00:12:28.433 "code": -22, 00:12:28.433 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:28.433 } 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:28.433 05:40:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:29.373 "name": "raid_bdev1", 00:12:29.373 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:29.373 "strip_size_kb": 0, 00:12:29.373 "state": "online", 00:12:29.373 "raid_level": "raid1", 00:12:29.373 "superblock": true, 00:12:29.373 "num_base_bdevs": 4, 00:12:29.373 "num_base_bdevs_discovered": 2, 00:12:29.373 "num_base_bdevs_operational": 2, 00:12:29.373 "base_bdevs_list": [ 00:12:29.373 { 00:12:29.373 "name": null, 00:12:29.373 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.373 "is_configured": false, 00:12:29.373 "data_offset": 0, 00:12:29.373 "data_size": 63488 00:12:29.373 }, 00:12:29.373 { 00:12:29.373 "name": null, 00:12:29.373 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.373 "is_configured": false, 00:12:29.373 "data_offset": 2048, 00:12:29.373 "data_size": 63488 00:12:29.373 }, 00:12:29.373 { 00:12:29.373 "name": "BaseBdev3", 00:12:29.373 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:29.373 "is_configured": true, 00:12:29.373 "data_offset": 2048, 00:12:29.373 "data_size": 63488 00:12:29.373 }, 00:12:29.373 { 00:12:29.373 "name": "BaseBdev4", 00:12:29.373 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:29.373 "is_configured": true, 00:12:29.373 "data_offset": 2048, 00:12:29.373 "data_size": 63488 00:12:29.373 } 00:12:29.373 ] 00:12:29.373 }' 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:29.373 05:40:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:29.943 "name": "raid_bdev1", 00:12:29.943 "uuid": "c4ef47ac-940a-45c0-90c3-f3887eab0f71", 00:12:29.943 "strip_size_kb": 0, 00:12:29.943 "state": "online", 00:12:29.943 "raid_level": "raid1", 00:12:29.943 "superblock": true, 00:12:29.943 "num_base_bdevs": 4, 00:12:29.943 "num_base_bdevs_discovered": 2, 00:12:29.943 "num_base_bdevs_operational": 2, 00:12:29.943 "base_bdevs_list": [ 00:12:29.943 { 00:12:29.943 "name": null, 00:12:29.943 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.943 "is_configured": false, 00:12:29.943 "data_offset": 0, 00:12:29.943 "data_size": 63488 00:12:29.943 }, 00:12:29.943 { 00:12:29.943 "name": null, 00:12:29.943 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.943 "is_configured": false, 00:12:29.943 "data_offset": 2048, 00:12:29.943 "data_size": 63488 00:12:29.943 }, 00:12:29.943 { 00:12:29.943 "name": "BaseBdev3", 00:12:29.943 "uuid": "a8f264e1-1641-5471-9ea4-daf7443cd986", 00:12:29.943 "is_configured": true, 00:12:29.943 "data_offset": 2048, 00:12:29.943 "data_size": 63488 00:12:29.943 }, 00:12:29.943 { 00:12:29.943 "name": "BaseBdev4", 00:12:29.943 "uuid": "0a2f0276-c5d4-50d3-a62e-8d52c1555754", 00:12:29.943 "is_configured": true, 00:12:29.943 "data_offset": 2048, 00:12:29.943 "data_size": 63488 00:12:29.943 } 00:12:29.943 ] 00:12:29.943 }' 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 88241 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 88241 ']' 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 88241 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 88241 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:29.943 killing process with pid 88241 00:12:29.943 Received shutdown signal, test time was about 60.000000 seconds 00:12:29.943 00:12:29.943 Latency(us) 00:12:29.943 [2024-12-07T05:40:03.311Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:29.943 [2024-12-07T05:40:03.311Z] =================================================================================================================== 00:12:29.943 [2024-12-07T05:40:03.311Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 88241' 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 88241 00:12:29.943 [2024-12-07 05:40:03.271577] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:29.943 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 88241 00:12:29.943 [2024-12-07 05:40:03.271748] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:29.944 [2024-12-07 05:40:03.271832] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:29.944 [2024-12-07 05:40:03.271847] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:12:30.203 [2024-12-07 05:40:03.371603] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:12:30.464 00:12:30.464 real 0m23.332s 00:12:30.464 user 0m28.688s 00:12:30.464 sys 0m3.596s 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:30.464 ************************************ 00:12:30.464 END TEST raid_rebuild_test_sb 00:12:30.464 ************************************ 00:12:30.464 05:40:03 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:12:30.464 05:40:03 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:30.464 05:40:03 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:30.464 05:40:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:30.464 ************************************ 00:12:30.464 START TEST raid_rebuild_test_io 00:12:30.464 ************************************ 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false true true 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=88979 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 88979 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 88979 ']' 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:30.464 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:30.464 05:40:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.724 [2024-12-07 05:40:03.865950] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:12:30.724 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:30.724 Zero copy mechanism will not be used. 00:12:30.724 [2024-12-07 05:40:03.866137] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88979 ] 00:12:30.724 [2024-12-07 05:40:04.001574] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:30.724 [2024-12-07 05:40:04.040029] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:30.984 [2024-12-07 05:40:04.117395] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:30.984 [2024-12-07 05:40:04.117437] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.555 BaseBdev1_malloc 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.555 [2024-12-07 05:40:04.729410] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:31.555 [2024-12-07 05:40:04.729477] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:31.555 [2024-12-07 05:40:04.729505] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:31.555 [2024-12-07 05:40:04.729517] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:31.555 [2024-12-07 05:40:04.731650] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:31.555 [2024-12-07 05:40:04.731683] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:31.555 BaseBdev1 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.555 BaseBdev2_malloc 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.555 [2024-12-07 05:40:04.753816] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:31.555 [2024-12-07 05:40:04.753869] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:31.555 [2024-12-07 05:40:04.753891] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:31.555 [2024-12-07 05:40:04.753899] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:31.555 [2024-12-07 05:40:04.755934] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:31.555 [2024-12-07 05:40:04.755975] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:31.555 BaseBdev2 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.555 BaseBdev3_malloc 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.555 [2024-12-07 05:40:04.782177] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:31.555 [2024-12-07 05:40:04.782229] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:31.555 [2024-12-07 05:40:04.782266] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:31.555 [2024-12-07 05:40:04.782274] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:31.555 [2024-12-07 05:40:04.784346] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:31.555 [2024-12-07 05:40:04.784382] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:31.555 BaseBdev3 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.555 BaseBdev4_malloc 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.555 [2024-12-07 05:40:04.829133] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:31.555 [2024-12-07 05:40:04.829270] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:31.555 [2024-12-07 05:40:04.829311] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:31.555 [2024-12-07 05:40:04.829326] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:31.555 [2024-12-07 05:40:04.832604] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:31.555 [2024-12-07 05:40:04.832670] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:31.555 BaseBdev4 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.555 spare_malloc 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.555 spare_delay 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.555 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.555 [2024-12-07 05:40:04.870052] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:31.556 [2024-12-07 05:40:04.870098] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:31.556 [2024-12-07 05:40:04.870119] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:31.556 [2024-12-07 05:40:04.870129] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:31.556 [2024-12-07 05:40:04.872403] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:31.556 [2024-12-07 05:40:04.872490] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:31.556 spare 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.556 [2024-12-07 05:40:04.882099] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:31.556 [2024-12-07 05:40:04.883985] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:31.556 [2024-12-07 05:40:04.884050] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:31.556 [2024-12-07 05:40:04.884099] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:31.556 [2024-12-07 05:40:04.884178] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:31.556 [2024-12-07 05:40:04.884195] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:31.556 [2024-12-07 05:40:04.884445] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:31.556 [2024-12-07 05:40:04.884584] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:31.556 [2024-12-07 05:40:04.884596] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:31.556 [2024-12-07 05:40:04.884746] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.556 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.816 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:31.816 "name": "raid_bdev1", 00:12:31.816 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:31.816 "strip_size_kb": 0, 00:12:31.816 "state": "online", 00:12:31.816 "raid_level": "raid1", 00:12:31.816 "superblock": false, 00:12:31.816 "num_base_bdevs": 4, 00:12:31.816 "num_base_bdevs_discovered": 4, 00:12:31.816 "num_base_bdevs_operational": 4, 00:12:31.816 "base_bdevs_list": [ 00:12:31.816 { 00:12:31.816 "name": "BaseBdev1", 00:12:31.816 "uuid": "46456178-2a43-524c-afe2-36f53f9cdbe1", 00:12:31.816 "is_configured": true, 00:12:31.816 "data_offset": 0, 00:12:31.816 "data_size": 65536 00:12:31.816 }, 00:12:31.816 { 00:12:31.816 "name": "BaseBdev2", 00:12:31.816 "uuid": "1f949f61-3f5b-5891-9707-574270aa95de", 00:12:31.816 "is_configured": true, 00:12:31.816 "data_offset": 0, 00:12:31.816 "data_size": 65536 00:12:31.816 }, 00:12:31.816 { 00:12:31.816 "name": "BaseBdev3", 00:12:31.816 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:31.816 "is_configured": true, 00:12:31.816 "data_offset": 0, 00:12:31.816 "data_size": 65536 00:12:31.816 }, 00:12:31.816 { 00:12:31.816 "name": "BaseBdev4", 00:12:31.816 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:31.816 "is_configured": true, 00:12:31.816 "data_offset": 0, 00:12:31.816 "data_size": 65536 00:12:31.816 } 00:12:31.816 ] 00:12:31.816 }' 00:12:31.816 05:40:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:31.816 05:40:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.084 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:32.084 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:32.084 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:32.084 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.084 [2024-12-07 05:40:05.385502] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:32.084 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:32.084 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:32.084 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:32.084 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:32.084 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:32.084 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.084 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.371 [2024-12-07 05:40:05.469034] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:32.371 "name": "raid_bdev1", 00:12:32.371 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:32.371 "strip_size_kb": 0, 00:12:32.371 "state": "online", 00:12:32.371 "raid_level": "raid1", 00:12:32.371 "superblock": false, 00:12:32.371 "num_base_bdevs": 4, 00:12:32.371 "num_base_bdevs_discovered": 3, 00:12:32.371 "num_base_bdevs_operational": 3, 00:12:32.371 "base_bdevs_list": [ 00:12:32.371 { 00:12:32.371 "name": null, 00:12:32.371 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:32.371 "is_configured": false, 00:12:32.371 "data_offset": 0, 00:12:32.371 "data_size": 65536 00:12:32.371 }, 00:12:32.371 { 00:12:32.371 "name": "BaseBdev2", 00:12:32.371 "uuid": "1f949f61-3f5b-5891-9707-574270aa95de", 00:12:32.371 "is_configured": true, 00:12:32.371 "data_offset": 0, 00:12:32.371 "data_size": 65536 00:12:32.371 }, 00:12:32.371 { 00:12:32.371 "name": "BaseBdev3", 00:12:32.371 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:32.371 "is_configured": true, 00:12:32.371 "data_offset": 0, 00:12:32.371 "data_size": 65536 00:12:32.371 }, 00:12:32.371 { 00:12:32.371 "name": "BaseBdev4", 00:12:32.371 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:32.371 "is_configured": true, 00:12:32.371 "data_offset": 0, 00:12:32.371 "data_size": 65536 00:12:32.371 } 00:12:32.371 ] 00:12:32.371 }' 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:32.371 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.371 [2024-12-07 05:40:05.538972] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:12:32.371 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:32.371 Zero copy mechanism will not be used. 00:12:32.371 Running I/O for 60 seconds... 00:12:32.649 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:32.649 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:32.649 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.649 [2024-12-07 05:40:05.889535] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:32.649 05:40:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:32.649 05:40:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:32.649 [2024-12-07 05:40:05.943576] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:12:32.649 [2024-12-07 05:40:05.945697] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:32.908 [2024-12-07 05:40:06.072798] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:32.908 [2024-12-07 05:40:06.073215] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:33.168 [2024-12-07 05:40:06.286938] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:33.168 [2024-12-07 05:40:06.287623] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:33.428 177.00 IOPS, 531.00 MiB/s [2024-12-07T05:40:06.796Z] [2024-12-07 05:40:06.628607] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:33.428 [2024-12-07 05:40:06.629035] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:33.428 [2024-12-07 05:40:06.765586] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:33.688 05:40:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:33.688 05:40:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:33.688 05:40:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:33.688 05:40:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:33.688 05:40:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:33.688 05:40:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.688 05:40:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.688 05:40:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:33.688 05:40:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.688 05:40:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:33.688 05:40:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:33.688 "name": "raid_bdev1", 00:12:33.688 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:33.688 "strip_size_kb": 0, 00:12:33.688 "state": "online", 00:12:33.688 "raid_level": "raid1", 00:12:33.688 "superblock": false, 00:12:33.688 "num_base_bdevs": 4, 00:12:33.688 "num_base_bdevs_discovered": 4, 00:12:33.688 "num_base_bdevs_operational": 4, 00:12:33.688 "process": { 00:12:33.688 "type": "rebuild", 00:12:33.688 "target": "spare", 00:12:33.688 "progress": { 00:12:33.688 "blocks": 12288, 00:12:33.688 "percent": 18 00:12:33.688 } 00:12:33.689 }, 00:12:33.689 "base_bdevs_list": [ 00:12:33.689 { 00:12:33.689 "name": "spare", 00:12:33.689 "uuid": "5272b9bd-7af6-53e1-ad7d-e50902a2267e", 00:12:33.689 "is_configured": true, 00:12:33.689 "data_offset": 0, 00:12:33.689 "data_size": 65536 00:12:33.689 }, 00:12:33.689 { 00:12:33.689 "name": "BaseBdev2", 00:12:33.689 "uuid": "1f949f61-3f5b-5891-9707-574270aa95de", 00:12:33.689 "is_configured": true, 00:12:33.689 "data_offset": 0, 00:12:33.689 "data_size": 65536 00:12:33.689 }, 00:12:33.689 { 00:12:33.689 "name": "BaseBdev3", 00:12:33.689 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:33.689 "is_configured": true, 00:12:33.689 "data_offset": 0, 00:12:33.689 "data_size": 65536 00:12:33.689 }, 00:12:33.689 { 00:12:33.689 "name": "BaseBdev4", 00:12:33.689 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:33.689 "is_configured": true, 00:12:33.689 "data_offset": 0, 00:12:33.689 "data_size": 65536 00:12:33.689 } 00:12:33.689 ] 00:12:33.689 }' 00:12:33.689 05:40:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:33.689 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:33.689 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.949 [2024-12-07 05:40:07.080566] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:33.949 [2024-12-07 05:40:07.145236] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:33.949 [2024-12-07 05:40:07.250069] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:33.949 [2024-12-07 05:40:07.252614] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:33.949 [2024-12-07 05:40:07.252711] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:33.949 [2024-12-07 05:40:07.252728] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:33.949 [2024-12-07 05:40:07.258939] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.949 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:33.950 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.950 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.209 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:34.209 "name": "raid_bdev1", 00:12:34.209 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:34.209 "strip_size_kb": 0, 00:12:34.210 "state": "online", 00:12:34.210 "raid_level": "raid1", 00:12:34.210 "superblock": false, 00:12:34.210 "num_base_bdevs": 4, 00:12:34.210 "num_base_bdevs_discovered": 3, 00:12:34.210 "num_base_bdevs_operational": 3, 00:12:34.210 "base_bdevs_list": [ 00:12:34.210 { 00:12:34.210 "name": null, 00:12:34.210 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.210 "is_configured": false, 00:12:34.210 "data_offset": 0, 00:12:34.210 "data_size": 65536 00:12:34.210 }, 00:12:34.210 { 00:12:34.210 "name": "BaseBdev2", 00:12:34.210 "uuid": "1f949f61-3f5b-5891-9707-574270aa95de", 00:12:34.210 "is_configured": true, 00:12:34.210 "data_offset": 0, 00:12:34.210 "data_size": 65536 00:12:34.210 }, 00:12:34.210 { 00:12:34.210 "name": "BaseBdev3", 00:12:34.210 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:34.210 "is_configured": true, 00:12:34.210 "data_offset": 0, 00:12:34.210 "data_size": 65536 00:12:34.210 }, 00:12:34.210 { 00:12:34.210 "name": "BaseBdev4", 00:12:34.210 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:34.210 "is_configured": true, 00:12:34.210 "data_offset": 0, 00:12:34.210 "data_size": 65536 00:12:34.210 } 00:12:34.210 ] 00:12:34.210 }' 00:12:34.210 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:34.210 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.470 158.00 IOPS, 474.00 MiB/s [2024-12-07T05:40:07.838Z] 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:34.470 "name": "raid_bdev1", 00:12:34.470 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:34.470 "strip_size_kb": 0, 00:12:34.470 "state": "online", 00:12:34.470 "raid_level": "raid1", 00:12:34.470 "superblock": false, 00:12:34.470 "num_base_bdevs": 4, 00:12:34.470 "num_base_bdevs_discovered": 3, 00:12:34.470 "num_base_bdevs_operational": 3, 00:12:34.470 "base_bdevs_list": [ 00:12:34.470 { 00:12:34.470 "name": null, 00:12:34.470 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.470 "is_configured": false, 00:12:34.470 "data_offset": 0, 00:12:34.470 "data_size": 65536 00:12:34.470 }, 00:12:34.470 { 00:12:34.470 "name": "BaseBdev2", 00:12:34.470 "uuid": "1f949f61-3f5b-5891-9707-574270aa95de", 00:12:34.470 "is_configured": true, 00:12:34.470 "data_offset": 0, 00:12:34.470 "data_size": 65536 00:12:34.470 }, 00:12:34.470 { 00:12:34.470 "name": "BaseBdev3", 00:12:34.470 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:34.470 "is_configured": true, 00:12:34.470 "data_offset": 0, 00:12:34.470 "data_size": 65536 00:12:34.470 }, 00:12:34.470 { 00:12:34.470 "name": "BaseBdev4", 00:12:34.470 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:34.470 "is_configured": true, 00:12:34.470 "data_offset": 0, 00:12:34.470 "data_size": 65536 00:12:34.470 } 00:12:34.470 ] 00:12:34.470 }' 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:34.470 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:34.471 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:34.471 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:34.471 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:34.471 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:34.471 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.471 [2024-12-07 05:40:07.834705] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:34.731 05:40:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.731 05:40:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:34.731 [2024-12-07 05:40:07.882496] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:12:34.731 [2024-12-07 05:40:07.884494] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:34.731 [2024-12-07 05:40:07.999981] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:34.731 [2024-12-07 05:40:08.001096] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:34.990 [2024-12-07 05:40:08.236633] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:34.990 [2024-12-07 05:40:08.237328] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:35.250 157.00 IOPS, 471.00 MiB/s [2024-12-07T05:40:08.618Z] [2024-12-07 05:40:08.588353] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:35.250 [2024-12-07 05:40:08.588905] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:35.511 [2024-12-07 05:40:08.697952] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:35.511 [2024-12-07 05:40:08.698314] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:35.511 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:35.511 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:35.511 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:35.511 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:35.511 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:35.511 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.511 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.511 05:40:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.511 05:40:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:35.771 "name": "raid_bdev1", 00:12:35.771 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:35.771 "strip_size_kb": 0, 00:12:35.771 "state": "online", 00:12:35.771 "raid_level": "raid1", 00:12:35.771 "superblock": false, 00:12:35.771 "num_base_bdevs": 4, 00:12:35.771 "num_base_bdevs_discovered": 4, 00:12:35.771 "num_base_bdevs_operational": 4, 00:12:35.771 "process": { 00:12:35.771 "type": "rebuild", 00:12:35.771 "target": "spare", 00:12:35.771 "progress": { 00:12:35.771 "blocks": 10240, 00:12:35.771 "percent": 15 00:12:35.771 } 00:12:35.771 }, 00:12:35.771 "base_bdevs_list": [ 00:12:35.771 { 00:12:35.771 "name": "spare", 00:12:35.771 "uuid": "5272b9bd-7af6-53e1-ad7d-e50902a2267e", 00:12:35.771 "is_configured": true, 00:12:35.771 "data_offset": 0, 00:12:35.771 "data_size": 65536 00:12:35.771 }, 00:12:35.771 { 00:12:35.771 "name": "BaseBdev2", 00:12:35.771 "uuid": "1f949f61-3f5b-5891-9707-574270aa95de", 00:12:35.771 "is_configured": true, 00:12:35.771 "data_offset": 0, 00:12:35.771 "data_size": 65536 00:12:35.771 }, 00:12:35.771 { 00:12:35.771 "name": "BaseBdev3", 00:12:35.771 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:35.771 "is_configured": true, 00:12:35.771 "data_offset": 0, 00:12:35.771 "data_size": 65536 00:12:35.771 }, 00:12:35.771 { 00:12:35.771 "name": "BaseBdev4", 00:12:35.771 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:35.771 "is_configured": true, 00:12:35.771 "data_offset": 0, 00:12:35.771 "data_size": 65536 00:12:35.771 } 00:12:35.771 ] 00:12:35.771 }' 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.771 05:40:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.771 [2024-12-07 05:40:09.000434] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:35.771 [2024-12-07 05:40:09.036384] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:35.771 [2024-12-07 05:40:09.047815] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:12:35.771 [2024-12-07 05:40:09.047880] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.771 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:35.771 "name": "raid_bdev1", 00:12:35.771 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:35.771 "strip_size_kb": 0, 00:12:35.771 "state": "online", 00:12:35.771 "raid_level": "raid1", 00:12:35.771 "superblock": false, 00:12:35.771 "num_base_bdevs": 4, 00:12:35.771 "num_base_bdevs_discovered": 3, 00:12:35.771 "num_base_bdevs_operational": 3, 00:12:35.771 "process": { 00:12:35.771 "type": "rebuild", 00:12:35.771 "target": "spare", 00:12:35.771 "progress": { 00:12:35.771 "blocks": 14336, 00:12:35.771 "percent": 21 00:12:35.771 } 00:12:35.771 }, 00:12:35.771 "base_bdevs_list": [ 00:12:35.771 { 00:12:35.771 "name": "spare", 00:12:35.771 "uuid": "5272b9bd-7af6-53e1-ad7d-e50902a2267e", 00:12:35.771 "is_configured": true, 00:12:35.771 "data_offset": 0, 00:12:35.771 "data_size": 65536 00:12:35.771 }, 00:12:35.771 { 00:12:35.771 "name": null, 00:12:35.771 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.772 "is_configured": false, 00:12:35.772 "data_offset": 0, 00:12:35.772 "data_size": 65536 00:12:35.772 }, 00:12:35.772 { 00:12:35.772 "name": "BaseBdev3", 00:12:35.772 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:35.772 "is_configured": true, 00:12:35.772 "data_offset": 0, 00:12:35.772 "data_size": 65536 00:12:35.772 }, 00:12:35.772 { 00:12:35.772 "name": "BaseBdev4", 00:12:35.772 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:35.772 "is_configured": true, 00:12:35.772 "data_offset": 0, 00:12:35.772 "data_size": 65536 00:12:35.772 } 00:12:35.772 ] 00:12:35.772 }' 00:12:35.772 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:35.772 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:35.772 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=386 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.031 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:36.031 [2024-12-07 05:40:09.177352] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:36.032 05:40:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.032 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:36.032 "name": "raid_bdev1", 00:12:36.032 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:36.032 "strip_size_kb": 0, 00:12:36.032 "state": "online", 00:12:36.032 "raid_level": "raid1", 00:12:36.032 "superblock": false, 00:12:36.032 "num_base_bdevs": 4, 00:12:36.032 "num_base_bdevs_discovered": 3, 00:12:36.032 "num_base_bdevs_operational": 3, 00:12:36.032 "process": { 00:12:36.032 "type": "rebuild", 00:12:36.032 "target": "spare", 00:12:36.032 "progress": { 00:12:36.032 "blocks": 14336, 00:12:36.032 "percent": 21 00:12:36.032 } 00:12:36.032 }, 00:12:36.032 "base_bdevs_list": [ 00:12:36.032 { 00:12:36.032 "name": "spare", 00:12:36.032 "uuid": "5272b9bd-7af6-53e1-ad7d-e50902a2267e", 00:12:36.032 "is_configured": true, 00:12:36.032 "data_offset": 0, 00:12:36.032 "data_size": 65536 00:12:36.032 }, 00:12:36.032 { 00:12:36.032 "name": null, 00:12:36.032 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.032 "is_configured": false, 00:12:36.032 "data_offset": 0, 00:12:36.032 "data_size": 65536 00:12:36.032 }, 00:12:36.032 { 00:12:36.032 "name": "BaseBdev3", 00:12:36.032 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:36.032 "is_configured": true, 00:12:36.032 "data_offset": 0, 00:12:36.032 "data_size": 65536 00:12:36.032 }, 00:12:36.032 { 00:12:36.032 "name": "BaseBdev4", 00:12:36.032 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:36.032 "is_configured": true, 00:12:36.032 "data_offset": 0, 00:12:36.032 "data_size": 65536 00:12:36.032 } 00:12:36.032 ] 00:12:36.032 }' 00:12:36.032 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:36.032 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:36.032 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:36.032 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:36.032 05:40:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:36.291 [2024-12-07 05:40:09.438595] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:36.861 137.00 IOPS, 411.00 MiB/s [2024-12-07T05:40:10.229Z] [2024-12-07 05:40:09.984302] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:37.121 "name": "raid_bdev1", 00:12:37.121 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:37.121 "strip_size_kb": 0, 00:12:37.121 "state": "online", 00:12:37.121 "raid_level": "raid1", 00:12:37.121 "superblock": false, 00:12:37.121 "num_base_bdevs": 4, 00:12:37.121 "num_base_bdevs_discovered": 3, 00:12:37.121 "num_base_bdevs_operational": 3, 00:12:37.121 "process": { 00:12:37.121 "type": "rebuild", 00:12:37.121 "target": "spare", 00:12:37.121 "progress": { 00:12:37.121 "blocks": 32768, 00:12:37.121 "percent": 50 00:12:37.121 } 00:12:37.121 }, 00:12:37.121 "base_bdevs_list": [ 00:12:37.121 { 00:12:37.121 "name": "spare", 00:12:37.121 "uuid": "5272b9bd-7af6-53e1-ad7d-e50902a2267e", 00:12:37.121 "is_configured": true, 00:12:37.121 "data_offset": 0, 00:12:37.121 "data_size": 65536 00:12:37.121 }, 00:12:37.121 { 00:12:37.121 "name": null, 00:12:37.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:37.121 "is_configured": false, 00:12:37.121 "data_offset": 0, 00:12:37.121 "data_size": 65536 00:12:37.121 }, 00:12:37.121 { 00:12:37.121 "name": "BaseBdev3", 00:12:37.121 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:37.121 "is_configured": true, 00:12:37.121 "data_offset": 0, 00:12:37.121 "data_size": 65536 00:12:37.121 }, 00:12:37.121 { 00:12:37.121 "name": "BaseBdev4", 00:12:37.121 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:37.121 "is_configured": true, 00:12:37.121 "data_offset": 0, 00:12:37.121 "data_size": 65536 00:12:37.121 } 00:12:37.121 ] 00:12:37.121 }' 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:37.121 05:40:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:37.381 [2024-12-07 05:40:10.535606] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:12:37.951 127.00 IOPS, 381.00 MiB/s [2024-12-07T05:40:11.319Z] [2024-12-07 05:40:11.089802] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:38.210 "name": "raid_bdev1", 00:12:38.210 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:38.210 "strip_size_kb": 0, 00:12:38.210 "state": "online", 00:12:38.210 "raid_level": "raid1", 00:12:38.210 "superblock": false, 00:12:38.210 "num_base_bdevs": 4, 00:12:38.210 "num_base_bdevs_discovered": 3, 00:12:38.210 "num_base_bdevs_operational": 3, 00:12:38.210 "process": { 00:12:38.210 "type": "rebuild", 00:12:38.210 "target": "spare", 00:12:38.210 "progress": { 00:12:38.210 "blocks": 51200, 00:12:38.210 "percent": 78 00:12:38.210 } 00:12:38.210 }, 00:12:38.210 "base_bdevs_list": [ 00:12:38.210 { 00:12:38.210 "name": "spare", 00:12:38.210 "uuid": "5272b9bd-7af6-53e1-ad7d-e50902a2267e", 00:12:38.210 "is_configured": true, 00:12:38.210 "data_offset": 0, 00:12:38.210 "data_size": 65536 00:12:38.210 }, 00:12:38.210 { 00:12:38.210 "name": null, 00:12:38.210 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.210 "is_configured": false, 00:12:38.210 "data_offset": 0, 00:12:38.210 "data_size": 65536 00:12:38.210 }, 00:12:38.210 { 00:12:38.210 "name": "BaseBdev3", 00:12:38.210 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:38.210 "is_configured": true, 00:12:38.210 "data_offset": 0, 00:12:38.210 "data_size": 65536 00:12:38.210 }, 00:12:38.210 { 00:12:38.210 "name": "BaseBdev4", 00:12:38.210 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:38.210 "is_configured": true, 00:12:38.210 "data_offset": 0, 00:12:38.210 "data_size": 65536 00:12:38.210 } 00:12:38.210 ] 00:12:38.210 }' 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:38.210 112.67 IOPS, 338.00 MiB/s [2024-12-07T05:40:11.578Z] 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:38.210 05:40:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:39.148 [2024-12-07 05:40:12.157056] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:39.148 [2024-12-07 05:40:12.261803] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:39.148 [2024-12-07 05:40:12.263939] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:39.409 101.29 IOPS, 303.86 MiB/s [2024-12-07T05:40:12.777Z] 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:39.409 "name": "raid_bdev1", 00:12:39.409 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:39.409 "strip_size_kb": 0, 00:12:39.409 "state": "online", 00:12:39.409 "raid_level": "raid1", 00:12:39.409 "superblock": false, 00:12:39.409 "num_base_bdevs": 4, 00:12:39.409 "num_base_bdevs_discovered": 3, 00:12:39.409 "num_base_bdevs_operational": 3, 00:12:39.409 "base_bdevs_list": [ 00:12:39.409 { 00:12:39.409 "name": "spare", 00:12:39.409 "uuid": "5272b9bd-7af6-53e1-ad7d-e50902a2267e", 00:12:39.409 "is_configured": true, 00:12:39.409 "data_offset": 0, 00:12:39.409 "data_size": 65536 00:12:39.409 }, 00:12:39.409 { 00:12:39.409 "name": null, 00:12:39.409 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:39.409 "is_configured": false, 00:12:39.409 "data_offset": 0, 00:12:39.409 "data_size": 65536 00:12:39.409 }, 00:12:39.409 { 00:12:39.409 "name": "BaseBdev3", 00:12:39.409 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:39.409 "is_configured": true, 00:12:39.409 "data_offset": 0, 00:12:39.409 "data_size": 65536 00:12:39.409 }, 00:12:39.409 { 00:12:39.409 "name": "BaseBdev4", 00:12:39.409 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:39.409 "is_configured": true, 00:12:39.409 "data_offset": 0, 00:12:39.409 "data_size": 65536 00:12:39.409 } 00:12:39.409 ] 00:12:39.409 }' 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:39.409 "name": "raid_bdev1", 00:12:39.409 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:39.409 "strip_size_kb": 0, 00:12:39.409 "state": "online", 00:12:39.409 "raid_level": "raid1", 00:12:39.409 "superblock": false, 00:12:39.409 "num_base_bdevs": 4, 00:12:39.409 "num_base_bdevs_discovered": 3, 00:12:39.409 "num_base_bdevs_operational": 3, 00:12:39.409 "base_bdevs_list": [ 00:12:39.409 { 00:12:39.409 "name": "spare", 00:12:39.409 "uuid": "5272b9bd-7af6-53e1-ad7d-e50902a2267e", 00:12:39.409 "is_configured": true, 00:12:39.409 "data_offset": 0, 00:12:39.409 "data_size": 65536 00:12:39.409 }, 00:12:39.409 { 00:12:39.409 "name": null, 00:12:39.409 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:39.409 "is_configured": false, 00:12:39.409 "data_offset": 0, 00:12:39.409 "data_size": 65536 00:12:39.409 }, 00:12:39.409 { 00:12:39.409 "name": "BaseBdev3", 00:12:39.409 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:39.409 "is_configured": true, 00:12:39.409 "data_offset": 0, 00:12:39.409 "data_size": 65536 00:12:39.409 }, 00:12:39.409 { 00:12:39.409 "name": "BaseBdev4", 00:12:39.409 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:39.409 "is_configured": true, 00:12:39.409 "data_offset": 0, 00:12:39.409 "data_size": 65536 00:12:39.409 } 00:12:39.409 ] 00:12:39.409 }' 00:12:39.409 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.668 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:39.668 "name": "raid_bdev1", 00:12:39.668 "uuid": "af35ba42-9769-45ff-b319-233a1f03c5bc", 00:12:39.668 "strip_size_kb": 0, 00:12:39.668 "state": "online", 00:12:39.668 "raid_level": "raid1", 00:12:39.668 "superblock": false, 00:12:39.668 "num_base_bdevs": 4, 00:12:39.668 "num_base_bdevs_discovered": 3, 00:12:39.668 "num_base_bdevs_operational": 3, 00:12:39.668 "base_bdevs_list": [ 00:12:39.668 { 00:12:39.668 "name": "spare", 00:12:39.668 "uuid": "5272b9bd-7af6-53e1-ad7d-e50902a2267e", 00:12:39.668 "is_configured": true, 00:12:39.668 "data_offset": 0, 00:12:39.668 "data_size": 65536 00:12:39.668 }, 00:12:39.668 { 00:12:39.668 "name": null, 00:12:39.668 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:39.668 "is_configured": false, 00:12:39.668 "data_offset": 0, 00:12:39.668 "data_size": 65536 00:12:39.668 }, 00:12:39.668 { 00:12:39.668 "name": "BaseBdev3", 00:12:39.668 "uuid": "a2ec004c-0bcc-5593-b294-eb7c785742ea", 00:12:39.668 "is_configured": true, 00:12:39.668 "data_offset": 0, 00:12:39.668 "data_size": 65536 00:12:39.668 }, 00:12:39.668 { 00:12:39.668 "name": "BaseBdev4", 00:12:39.668 "uuid": "cf856c1c-1ec8-5382-a96d-964eda051dc7", 00:12:39.668 "is_configured": true, 00:12:39.668 "data_offset": 0, 00:12:39.668 "data_size": 65536 00:12:39.668 } 00:12:39.668 ] 00:12:39.668 }' 00:12:39.669 05:40:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:39.669 05:40:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.926 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:39.926 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.926 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.926 [2024-12-07 05:40:13.229983] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:39.926 [2024-12-07 05:40:13.230062] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:40.184 00:12:40.184 Latency(us) 00:12:40.184 [2024-12-07T05:40:13.552Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:40.184 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:40.184 raid_bdev1 : 7.80 94.70 284.11 0.00 0.00 14838.81 284.39 115389.15 00:12:40.184 [2024-12-07T05:40:13.552Z] =================================================================================================================== 00:12:40.184 [2024-12-07T05:40:13.552Z] Total : 94.70 284.11 0.00 0.00 14838.81 284.39 115389.15 00:12:40.184 [2024-12-07 05:40:13.332989] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:40.184 [2024-12-07 05:40:13.333072] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:40.184 [2024-12-07 05:40:13.333193] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:40.184 [2024-12-07 05:40:13.333236] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:40.184 { 00:12:40.184 "results": [ 00:12:40.184 { 00:12:40.184 "job": "raid_bdev1", 00:12:40.184 "core_mask": "0x1", 00:12:40.184 "workload": "randrw", 00:12:40.184 "percentage": 50, 00:12:40.184 "status": "finished", 00:12:40.184 "queue_depth": 2, 00:12:40.184 "io_size": 3145728, 00:12:40.184 "runtime": 7.803404, 00:12:40.184 "iops": 94.70226070571253, 00:12:40.184 "mibps": 284.1067821171376, 00:12:40.184 "io_failed": 0, 00:12:40.184 "io_timeout": 0, 00:12:40.184 "avg_latency_us": 14838.806866354274, 00:12:40.184 "min_latency_us": 284.3947598253275, 00:12:40.184 "max_latency_us": 115389.14934497817 00:12:40.184 } 00:12:40.184 ], 00:12:40.184 "core_count": 1 00:12:40.184 } 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:40.184 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:40.443 /dev/nbd0 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:40.443 1+0 records in 00:12:40.443 1+0 records out 00:12:40.443 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000336904 s, 12.2 MB/s 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:40.443 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:12:40.701 /dev/nbd1 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:40.701 1+0 records in 00:12:40.701 1+0 records out 00:12:40.701 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000454164 s, 9.0 MB/s 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:40.701 05:40:13 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:40.960 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:12:41.219 /dev/nbd1 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:41.219 1+0 records in 00:12:41.219 1+0 records out 00:12:41.219 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000518876 s, 7.9 MB/s 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:41.219 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 88979 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 88979 ']' 00:12:41.479 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 88979 00:12:41.739 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:12:41.739 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:41.739 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 88979 00:12:41.739 killing process with pid 88979 00:12:41.739 Received shutdown signal, test time was about 9.358793 seconds 00:12:41.739 00:12:41.739 Latency(us) 00:12:41.739 [2024-12-07T05:40:15.107Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:41.739 [2024-12-07T05:40:15.107Z] =================================================================================================================== 00:12:41.739 [2024-12-07T05:40:15.107Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:41.739 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:41.739 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:41.739 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 88979' 00:12:41.739 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 88979 00:12:41.739 [2024-12-07 05:40:14.882142] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:41.739 05:40:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 88979 00:12:41.739 [2024-12-07 05:40:14.927656] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:41.998 00:12:41.998 real 0m11.350s 00:12:41.998 user 0m14.655s 00:12:41.998 sys 0m1.653s 00:12:41.998 ************************************ 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:41.998 END TEST raid_rebuild_test_io 00:12:41.998 ************************************ 00:12:41.998 05:40:15 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:12:41.998 05:40:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:41.998 05:40:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:41.998 05:40:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:41.998 ************************************ 00:12:41.998 START TEST raid_rebuild_test_sb_io 00:12:41.998 ************************************ 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true true true 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89366 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89366 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 89366 ']' 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:41.998 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:41.998 05:40:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:41.998 [2024-12-07 05:40:15.272223] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:12:41.998 [2024-12-07 05:40:15.272409] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:12:41.998 Zero copy mechanism will not be used. 00:12:41.998 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89366 ] 00:12:42.256 [2024-12-07 05:40:15.426315] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:42.256 [2024-12-07 05:40:15.451005] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:42.256 [2024-12-07 05:40:15.493463] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:42.256 [2024-12-07 05:40:15.493580] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.824 BaseBdev1_malloc 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.824 [2024-12-07 05:40:16.120956] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:42.824 [2024-12-07 05:40:16.121012] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:42.824 [2024-12-07 05:40:16.121037] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:42.824 [2024-12-07 05:40:16.121048] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:42.824 [2024-12-07 05:40:16.123080] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:42.824 [2024-12-07 05:40:16.123118] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:42.824 BaseBdev1 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.824 BaseBdev2_malloc 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.824 [2024-12-07 05:40:16.149386] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:42.824 [2024-12-07 05:40:16.149440] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:42.824 [2024-12-07 05:40:16.149478] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:42.824 [2024-12-07 05:40:16.149486] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:42.824 [2024-12-07 05:40:16.151600] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:42.824 [2024-12-07 05:40:16.151733] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:42.824 BaseBdev2 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.824 BaseBdev3_malloc 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:42.824 [2024-12-07 05:40:16.177859] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:42.824 [2024-12-07 05:40:16.177969] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:42.824 [2024-12-07 05:40:16.177995] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:42.824 [2024-12-07 05:40:16.178005] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:42.824 [2024-12-07 05:40:16.180159] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:42.824 [2024-12-07 05:40:16.180195] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:42.824 BaseBdev3 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.824 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.084 BaseBdev4_malloc 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.084 [2024-12-07 05:40:16.214039] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:43.084 [2024-12-07 05:40:16.214087] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:43.084 [2024-12-07 05:40:16.214108] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:43.084 [2024-12-07 05:40:16.214116] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:43.084 [2024-12-07 05:40:16.216227] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:43.084 [2024-12-07 05:40:16.216261] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:43.084 BaseBdev4 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.084 spare_malloc 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.084 spare_delay 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.084 [2024-12-07 05:40:16.246497] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:43.084 [2024-12-07 05:40:16.246543] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:43.084 [2024-12-07 05:40:16.246576] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:43.084 [2024-12-07 05:40:16.246585] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:43.084 [2024-12-07 05:40:16.248703] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:43.084 [2024-12-07 05:40:16.248735] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:43.084 spare 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.084 [2024-12-07 05:40:16.258582] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:43.084 [2024-12-07 05:40:16.260463] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:43.084 [2024-12-07 05:40:16.260588] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:43.084 [2024-12-07 05:40:16.260665] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:43.084 [2024-12-07 05:40:16.260854] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:43.084 [2024-12-07 05:40:16.260871] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:43.084 [2024-12-07 05:40:16.261100] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:43.084 [2024-12-07 05:40:16.261237] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:43.084 [2024-12-07 05:40:16.261249] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:43.084 [2024-12-07 05:40:16.261364] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:43.084 "name": "raid_bdev1", 00:12:43.084 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:43.084 "strip_size_kb": 0, 00:12:43.084 "state": "online", 00:12:43.084 "raid_level": "raid1", 00:12:43.084 "superblock": true, 00:12:43.084 "num_base_bdevs": 4, 00:12:43.084 "num_base_bdevs_discovered": 4, 00:12:43.084 "num_base_bdevs_operational": 4, 00:12:43.084 "base_bdevs_list": [ 00:12:43.084 { 00:12:43.084 "name": "BaseBdev1", 00:12:43.084 "uuid": "fec53896-3959-56d9-8ecb-7f67af99dd35", 00:12:43.084 "is_configured": true, 00:12:43.084 "data_offset": 2048, 00:12:43.084 "data_size": 63488 00:12:43.084 }, 00:12:43.084 { 00:12:43.084 "name": "BaseBdev2", 00:12:43.084 "uuid": "a7a1cd66-80be-5b03-a7da-a02afc706ab6", 00:12:43.084 "is_configured": true, 00:12:43.084 "data_offset": 2048, 00:12:43.084 "data_size": 63488 00:12:43.084 }, 00:12:43.084 { 00:12:43.084 "name": "BaseBdev3", 00:12:43.084 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:43.084 "is_configured": true, 00:12:43.084 "data_offset": 2048, 00:12:43.084 "data_size": 63488 00:12:43.084 }, 00:12:43.084 { 00:12:43.084 "name": "BaseBdev4", 00:12:43.084 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:43.084 "is_configured": true, 00:12:43.084 "data_offset": 2048, 00:12:43.084 "data_size": 63488 00:12:43.084 } 00:12:43.084 ] 00:12:43.084 }' 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:43.084 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.652 [2024-12-07 05:40:16.757977] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.652 [2024-12-07 05:40:16.829516] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:43.652 "name": "raid_bdev1", 00:12:43.652 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:43.652 "strip_size_kb": 0, 00:12:43.652 "state": "online", 00:12:43.652 "raid_level": "raid1", 00:12:43.652 "superblock": true, 00:12:43.652 "num_base_bdevs": 4, 00:12:43.652 "num_base_bdevs_discovered": 3, 00:12:43.652 "num_base_bdevs_operational": 3, 00:12:43.652 "base_bdevs_list": [ 00:12:43.652 { 00:12:43.652 "name": null, 00:12:43.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:43.652 "is_configured": false, 00:12:43.652 "data_offset": 0, 00:12:43.652 "data_size": 63488 00:12:43.652 }, 00:12:43.652 { 00:12:43.652 "name": "BaseBdev2", 00:12:43.652 "uuid": "a7a1cd66-80be-5b03-a7da-a02afc706ab6", 00:12:43.652 "is_configured": true, 00:12:43.652 "data_offset": 2048, 00:12:43.652 "data_size": 63488 00:12:43.652 }, 00:12:43.652 { 00:12:43.652 "name": "BaseBdev3", 00:12:43.652 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:43.652 "is_configured": true, 00:12:43.652 "data_offset": 2048, 00:12:43.652 "data_size": 63488 00:12:43.652 }, 00:12:43.652 { 00:12:43.652 "name": "BaseBdev4", 00:12:43.652 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:43.652 "is_configured": true, 00:12:43.652 "data_offset": 2048, 00:12:43.652 "data_size": 63488 00:12:43.652 } 00:12:43.652 ] 00:12:43.652 }' 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:43.652 05:40:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.652 [2024-12-07 05:40:16.903413] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:12:43.652 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:43.652 Zero copy mechanism will not be used. 00:12:43.652 Running I/O for 60 seconds... 00:12:43.912 05:40:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:43.912 05:40:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.912 05:40:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.912 [2024-12-07 05:40:17.271769] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:44.171 05:40:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.171 05:40:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:44.171 [2024-12-07 05:40:17.325713] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:12:44.171 [2024-12-07 05:40:17.327798] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:44.171 [2024-12-07 05:40:17.435720] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:44.171 [2024-12-07 05:40:17.436319] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:44.431 [2024-12-07 05:40:17.652480] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:44.432 [2024-12-07 05:40:17.652808] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:44.692 139.00 IOPS, 417.00 MiB/s [2024-12-07T05:40:18.060Z] [2024-12-07 05:40:17.975159] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:44.952 [2024-12-07 05:40:18.084031] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:44.952 [2024-12-07 05:40:18.084583] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:44.952 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:44.952 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:44.952 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:44.952 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:44.952 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:44.952 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.952 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.952 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.952 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:45.213 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.213 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:45.213 "name": "raid_bdev1", 00:12:45.213 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:45.213 "strip_size_kb": 0, 00:12:45.213 "state": "online", 00:12:45.213 "raid_level": "raid1", 00:12:45.213 "superblock": true, 00:12:45.213 "num_base_bdevs": 4, 00:12:45.213 "num_base_bdevs_discovered": 4, 00:12:45.213 "num_base_bdevs_operational": 4, 00:12:45.213 "process": { 00:12:45.213 "type": "rebuild", 00:12:45.213 "target": "spare", 00:12:45.213 "progress": { 00:12:45.213 "blocks": 12288, 00:12:45.213 "percent": 19 00:12:45.213 } 00:12:45.213 }, 00:12:45.213 "base_bdevs_list": [ 00:12:45.213 { 00:12:45.213 "name": "spare", 00:12:45.213 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:45.213 "is_configured": true, 00:12:45.213 "data_offset": 2048, 00:12:45.213 "data_size": 63488 00:12:45.213 }, 00:12:45.213 { 00:12:45.213 "name": "BaseBdev2", 00:12:45.213 "uuid": "a7a1cd66-80be-5b03-a7da-a02afc706ab6", 00:12:45.213 "is_configured": true, 00:12:45.213 "data_offset": 2048, 00:12:45.213 "data_size": 63488 00:12:45.213 }, 00:12:45.213 { 00:12:45.213 "name": "BaseBdev3", 00:12:45.213 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:45.213 "is_configured": true, 00:12:45.213 "data_offset": 2048, 00:12:45.213 "data_size": 63488 00:12:45.213 }, 00:12:45.213 { 00:12:45.213 "name": "BaseBdev4", 00:12:45.213 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:45.213 "is_configured": true, 00:12:45.213 "data_offset": 2048, 00:12:45.213 "data_size": 63488 00:12:45.213 } 00:12:45.213 ] 00:12:45.213 }' 00:12:45.213 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:45.213 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:45.213 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:45.213 [2024-12-07 05:40:18.415582] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:45.213 [2024-12-07 05:40:18.416092] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:45.213 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:45.213 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:45.213 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.213 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:45.213 [2024-12-07 05:40:18.459278] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:45.213 [2024-12-07 05:40:18.525346] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:45.213 [2024-12-07 05:40:18.532129] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:45.473 [2024-12-07 05:40:18.634056] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:45.473 [2024-12-07 05:40:18.642721] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:45.473 [2024-12-07 05:40:18.642811] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:45.473 [2024-12-07 05:40:18.642843] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:45.473 [2024-12-07 05:40:18.671708] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:45.473 "name": "raid_bdev1", 00:12:45.473 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:45.473 "strip_size_kb": 0, 00:12:45.473 "state": "online", 00:12:45.473 "raid_level": "raid1", 00:12:45.473 "superblock": true, 00:12:45.473 "num_base_bdevs": 4, 00:12:45.473 "num_base_bdevs_discovered": 3, 00:12:45.473 "num_base_bdevs_operational": 3, 00:12:45.473 "base_bdevs_list": [ 00:12:45.473 { 00:12:45.473 "name": null, 00:12:45.473 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:45.473 "is_configured": false, 00:12:45.473 "data_offset": 0, 00:12:45.473 "data_size": 63488 00:12:45.473 }, 00:12:45.473 { 00:12:45.473 "name": "BaseBdev2", 00:12:45.473 "uuid": "a7a1cd66-80be-5b03-a7da-a02afc706ab6", 00:12:45.473 "is_configured": true, 00:12:45.473 "data_offset": 2048, 00:12:45.473 "data_size": 63488 00:12:45.473 }, 00:12:45.473 { 00:12:45.473 "name": "BaseBdev3", 00:12:45.473 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:45.473 "is_configured": true, 00:12:45.473 "data_offset": 2048, 00:12:45.473 "data_size": 63488 00:12:45.473 }, 00:12:45.473 { 00:12:45.473 "name": "BaseBdev4", 00:12:45.473 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:45.473 "is_configured": true, 00:12:45.473 "data_offset": 2048, 00:12:45.473 "data_size": 63488 00:12:45.473 } 00:12:45.473 ] 00:12:45.473 }' 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:45.473 05:40:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:45.992 120.00 IOPS, 360.00 MiB/s [2024-12-07T05:40:19.360Z] 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:45.992 "name": "raid_bdev1", 00:12:45.992 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:45.992 "strip_size_kb": 0, 00:12:45.992 "state": "online", 00:12:45.992 "raid_level": "raid1", 00:12:45.992 "superblock": true, 00:12:45.992 "num_base_bdevs": 4, 00:12:45.992 "num_base_bdevs_discovered": 3, 00:12:45.992 "num_base_bdevs_operational": 3, 00:12:45.992 "base_bdevs_list": [ 00:12:45.992 { 00:12:45.992 "name": null, 00:12:45.992 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:45.992 "is_configured": false, 00:12:45.992 "data_offset": 0, 00:12:45.992 "data_size": 63488 00:12:45.992 }, 00:12:45.992 { 00:12:45.992 "name": "BaseBdev2", 00:12:45.992 "uuid": "a7a1cd66-80be-5b03-a7da-a02afc706ab6", 00:12:45.992 "is_configured": true, 00:12:45.992 "data_offset": 2048, 00:12:45.992 "data_size": 63488 00:12:45.992 }, 00:12:45.992 { 00:12:45.992 "name": "BaseBdev3", 00:12:45.992 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:45.992 "is_configured": true, 00:12:45.992 "data_offset": 2048, 00:12:45.992 "data_size": 63488 00:12:45.992 }, 00:12:45.992 { 00:12:45.992 "name": "BaseBdev4", 00:12:45.992 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:45.992 "is_configured": true, 00:12:45.992 "data_offset": 2048, 00:12:45.992 "data_size": 63488 00:12:45.992 } 00:12:45.992 ] 00:12:45.992 }' 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:45.992 [2024-12-07 05:40:19.279125] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.992 05:40:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:45.992 [2024-12-07 05:40:19.315229] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:12:45.992 [2024-12-07 05:40:19.317230] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:46.252 [2024-12-07 05:40:19.432287] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:46.252 [2024-12-07 05:40:19.433705] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:46.511 [2024-12-07 05:40:19.642591] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:46.511 [2024-12-07 05:40:19.643271] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:46.770 141.67 IOPS, 425.00 MiB/s [2024-12-07T05:40:20.138Z] [2024-12-07 05:40:19.973675] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:46.770 [2024-12-07 05:40:19.974152] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:47.030 [2024-12-07 05:40:20.176409] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:47.030 [2024-12-07 05:40:20.177035] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:47.030 "name": "raid_bdev1", 00:12:47.030 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:47.030 "strip_size_kb": 0, 00:12:47.030 "state": "online", 00:12:47.030 "raid_level": "raid1", 00:12:47.030 "superblock": true, 00:12:47.030 "num_base_bdevs": 4, 00:12:47.030 "num_base_bdevs_discovered": 4, 00:12:47.030 "num_base_bdevs_operational": 4, 00:12:47.030 "process": { 00:12:47.030 "type": "rebuild", 00:12:47.030 "target": "spare", 00:12:47.030 "progress": { 00:12:47.030 "blocks": 10240, 00:12:47.030 "percent": 16 00:12:47.030 } 00:12:47.030 }, 00:12:47.030 "base_bdevs_list": [ 00:12:47.030 { 00:12:47.030 "name": "spare", 00:12:47.030 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:47.030 "is_configured": true, 00:12:47.030 "data_offset": 2048, 00:12:47.030 "data_size": 63488 00:12:47.030 }, 00:12:47.030 { 00:12:47.030 "name": "BaseBdev2", 00:12:47.030 "uuid": "a7a1cd66-80be-5b03-a7da-a02afc706ab6", 00:12:47.030 "is_configured": true, 00:12:47.030 "data_offset": 2048, 00:12:47.030 "data_size": 63488 00:12:47.030 }, 00:12:47.030 { 00:12:47.030 "name": "BaseBdev3", 00:12:47.030 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:47.030 "is_configured": true, 00:12:47.030 "data_offset": 2048, 00:12:47.030 "data_size": 63488 00:12:47.030 }, 00:12:47.030 { 00:12:47.030 "name": "BaseBdev4", 00:12:47.030 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:47.030 "is_configured": true, 00:12:47.030 "data_offset": 2048, 00:12:47.030 "data_size": 63488 00:12:47.030 } 00:12:47.030 ] 00:12:47.030 }' 00:12:47.030 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:47.289 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:47.289 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:47.289 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:47.289 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:47.289 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:47.289 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:47.289 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:47.289 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:47.289 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:47.289 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:47.289 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.289 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.289 [2024-12-07 05:40:20.461224] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:47.289 [2024-12-07 05:40:20.508958] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:47.556 [2024-12-07 05:40:20.710681] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:12:47.556 [2024-12-07 05:40:20.710755] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:12:47.556 [2024-12-07 05:40:20.710822] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:47.556 "name": "raid_bdev1", 00:12:47.556 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:47.556 "strip_size_kb": 0, 00:12:47.556 "state": "online", 00:12:47.556 "raid_level": "raid1", 00:12:47.556 "superblock": true, 00:12:47.556 "num_base_bdevs": 4, 00:12:47.556 "num_base_bdevs_discovered": 3, 00:12:47.556 "num_base_bdevs_operational": 3, 00:12:47.556 "process": { 00:12:47.556 "type": "rebuild", 00:12:47.556 "target": "spare", 00:12:47.556 "progress": { 00:12:47.556 "blocks": 14336, 00:12:47.556 "percent": 22 00:12:47.556 } 00:12:47.556 }, 00:12:47.556 "base_bdevs_list": [ 00:12:47.556 { 00:12:47.556 "name": "spare", 00:12:47.556 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:47.556 "is_configured": true, 00:12:47.556 "data_offset": 2048, 00:12:47.556 "data_size": 63488 00:12:47.556 }, 00:12:47.556 { 00:12:47.556 "name": null, 00:12:47.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:47.556 "is_configured": false, 00:12:47.556 "data_offset": 0, 00:12:47.556 "data_size": 63488 00:12:47.556 }, 00:12:47.556 { 00:12:47.556 "name": "BaseBdev3", 00:12:47.556 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:47.556 "is_configured": true, 00:12:47.556 "data_offset": 2048, 00:12:47.556 "data_size": 63488 00:12:47.556 }, 00:12:47.556 { 00:12:47.556 "name": "BaseBdev4", 00:12:47.556 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:47.556 "is_configured": true, 00:12:47.556 "data_offset": 2048, 00:12:47.556 "data_size": 63488 00:12:47.556 } 00:12:47.556 ] 00:12:47.556 }' 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=397 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.556 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:47.556 "name": "raid_bdev1", 00:12:47.556 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:47.556 "strip_size_kb": 0, 00:12:47.556 "state": "online", 00:12:47.556 "raid_level": "raid1", 00:12:47.556 "superblock": true, 00:12:47.556 "num_base_bdevs": 4, 00:12:47.556 "num_base_bdevs_discovered": 3, 00:12:47.556 "num_base_bdevs_operational": 3, 00:12:47.556 "process": { 00:12:47.556 "type": "rebuild", 00:12:47.556 "target": "spare", 00:12:47.556 "progress": { 00:12:47.556 "blocks": 14336, 00:12:47.556 "percent": 22 00:12:47.556 } 00:12:47.556 }, 00:12:47.556 "base_bdevs_list": [ 00:12:47.556 { 00:12:47.556 "name": "spare", 00:12:47.556 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:47.556 "is_configured": true, 00:12:47.556 "data_offset": 2048, 00:12:47.556 "data_size": 63488 00:12:47.556 }, 00:12:47.556 { 00:12:47.556 "name": null, 00:12:47.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:47.556 "is_configured": false, 00:12:47.556 "data_offset": 0, 00:12:47.556 "data_size": 63488 00:12:47.556 }, 00:12:47.556 { 00:12:47.556 "name": "BaseBdev3", 00:12:47.556 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:47.556 "is_configured": true, 00:12:47.556 "data_offset": 2048, 00:12:47.557 "data_size": 63488 00:12:47.557 }, 00:12:47.557 { 00:12:47.557 "name": "BaseBdev4", 00:12:47.557 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:47.557 "is_configured": true, 00:12:47.557 "data_offset": 2048, 00:12:47.557 "data_size": 63488 00:12:47.557 } 00:12:47.557 ] 00:12:47.557 }' 00:12:47.557 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:47.557 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:47.557 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:47.834 121.75 IOPS, 365.25 MiB/s [2024-12-07T05:40:21.202Z] 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:47.834 05:40:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:47.834 [2024-12-07 05:40:20.951636] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:47.834 [2024-12-07 05:40:21.193670] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:48.092 [2024-12-07 05:40:21.407154] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:48.092 [2024-12-07 05:40:21.407460] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:48.660 [2024-12-07 05:40:21.739118] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:12:48.660 [2024-12-07 05:40:21.863803] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:12:48.660 107.20 IOPS, 321.60 MiB/s [2024-12-07T05:40:22.028Z] 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:48.660 "name": "raid_bdev1", 00:12:48.660 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:48.660 "strip_size_kb": 0, 00:12:48.660 "state": "online", 00:12:48.660 "raid_level": "raid1", 00:12:48.660 "superblock": true, 00:12:48.660 "num_base_bdevs": 4, 00:12:48.660 "num_base_bdevs_discovered": 3, 00:12:48.660 "num_base_bdevs_operational": 3, 00:12:48.660 "process": { 00:12:48.660 "type": "rebuild", 00:12:48.660 "target": "spare", 00:12:48.660 "progress": { 00:12:48.660 "blocks": 28672, 00:12:48.660 "percent": 45 00:12:48.660 } 00:12:48.660 }, 00:12:48.660 "base_bdevs_list": [ 00:12:48.660 { 00:12:48.660 "name": "spare", 00:12:48.660 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:48.660 "is_configured": true, 00:12:48.660 "data_offset": 2048, 00:12:48.660 "data_size": 63488 00:12:48.660 }, 00:12:48.660 { 00:12:48.660 "name": null, 00:12:48.660 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.660 "is_configured": false, 00:12:48.660 "data_offset": 0, 00:12:48.660 "data_size": 63488 00:12:48.660 }, 00:12:48.660 { 00:12:48.660 "name": "BaseBdev3", 00:12:48.660 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:48.660 "is_configured": true, 00:12:48.660 "data_offset": 2048, 00:12:48.660 "data_size": 63488 00:12:48.660 }, 00:12:48.660 { 00:12:48.660 "name": "BaseBdev4", 00:12:48.660 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:48.660 "is_configured": true, 00:12:48.660 "data_offset": 2048, 00:12:48.660 "data_size": 63488 00:12:48.660 } 00:12:48.660 ] 00:12:48.660 }' 00:12:48.660 05:40:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:48.660 05:40:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:48.660 05:40:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:48.919 05:40:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:48.919 05:40:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:49.856 [2024-12-07 05:40:22.875628] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:12:49.856 99.17 IOPS, 297.50 MiB/s [2024-12-07T05:40:23.224Z] [2024-12-07 05:40:22.988619] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:49.856 "name": "raid_bdev1", 00:12:49.856 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:49.856 "strip_size_kb": 0, 00:12:49.856 "state": "online", 00:12:49.856 "raid_level": "raid1", 00:12:49.856 "superblock": true, 00:12:49.856 "num_base_bdevs": 4, 00:12:49.856 "num_base_bdevs_discovered": 3, 00:12:49.856 "num_base_bdevs_operational": 3, 00:12:49.856 "process": { 00:12:49.856 "type": "rebuild", 00:12:49.856 "target": "spare", 00:12:49.856 "progress": { 00:12:49.856 "blocks": 47104, 00:12:49.856 "percent": 74 00:12:49.856 } 00:12:49.856 }, 00:12:49.856 "base_bdevs_list": [ 00:12:49.856 { 00:12:49.856 "name": "spare", 00:12:49.856 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:49.856 "is_configured": true, 00:12:49.856 "data_offset": 2048, 00:12:49.856 "data_size": 63488 00:12:49.856 }, 00:12:49.856 { 00:12:49.856 "name": null, 00:12:49.856 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:49.856 "is_configured": false, 00:12:49.856 "data_offset": 0, 00:12:49.856 "data_size": 63488 00:12:49.856 }, 00:12:49.856 { 00:12:49.856 "name": "BaseBdev3", 00:12:49.856 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:49.856 "is_configured": true, 00:12:49.856 "data_offset": 2048, 00:12:49.856 "data_size": 63488 00:12:49.856 }, 00:12:49.856 { 00:12:49.856 "name": "BaseBdev4", 00:12:49.856 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:49.856 "is_configured": true, 00:12:49.856 "data_offset": 2048, 00:12:49.856 "data_size": 63488 00:12:49.856 } 00:12:49.856 ] 00:12:49.856 }' 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:49.856 05:40:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:50.425 [2024-12-07 05:40:23.644654] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:12:50.685 [2024-12-07 05:40:23.873276] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:50.685 89.29 IOPS, 267.86 MiB/s [2024-12-07T05:40:24.053Z] [2024-12-07 05:40:23.978274] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:50.685 [2024-12-07 05:40:23.981654] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.945 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:50.945 "name": "raid_bdev1", 00:12:50.945 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:50.945 "strip_size_kb": 0, 00:12:50.945 "state": "online", 00:12:50.945 "raid_level": "raid1", 00:12:50.945 "superblock": true, 00:12:50.945 "num_base_bdevs": 4, 00:12:50.945 "num_base_bdevs_discovered": 3, 00:12:50.945 "num_base_bdevs_operational": 3, 00:12:50.945 "base_bdevs_list": [ 00:12:50.945 { 00:12:50.945 "name": "spare", 00:12:50.945 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:50.945 "is_configured": true, 00:12:50.945 "data_offset": 2048, 00:12:50.945 "data_size": 63488 00:12:50.945 }, 00:12:50.945 { 00:12:50.945 "name": null, 00:12:50.945 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:50.945 "is_configured": false, 00:12:50.945 "data_offset": 0, 00:12:50.945 "data_size": 63488 00:12:50.945 }, 00:12:50.945 { 00:12:50.945 "name": "BaseBdev3", 00:12:50.945 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:50.945 "is_configured": true, 00:12:50.945 "data_offset": 2048, 00:12:50.945 "data_size": 63488 00:12:50.945 }, 00:12:50.945 { 00:12:50.945 "name": "BaseBdev4", 00:12:50.945 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:50.945 "is_configured": true, 00:12:50.945 "data_offset": 2048, 00:12:50.945 "data_size": 63488 00:12:50.945 } 00:12:50.945 ] 00:12:50.945 }' 00:12:50.946 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:51.206 "name": "raid_bdev1", 00:12:51.206 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:51.206 "strip_size_kb": 0, 00:12:51.206 "state": "online", 00:12:51.206 "raid_level": "raid1", 00:12:51.206 "superblock": true, 00:12:51.206 "num_base_bdevs": 4, 00:12:51.206 "num_base_bdevs_discovered": 3, 00:12:51.206 "num_base_bdevs_operational": 3, 00:12:51.206 "base_bdevs_list": [ 00:12:51.206 { 00:12:51.206 "name": "spare", 00:12:51.206 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:51.206 "is_configured": true, 00:12:51.206 "data_offset": 2048, 00:12:51.206 "data_size": 63488 00:12:51.206 }, 00:12:51.206 { 00:12:51.206 "name": null, 00:12:51.206 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:51.206 "is_configured": false, 00:12:51.206 "data_offset": 0, 00:12:51.206 "data_size": 63488 00:12:51.206 }, 00:12:51.206 { 00:12:51.206 "name": "BaseBdev3", 00:12:51.206 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:51.206 "is_configured": true, 00:12:51.206 "data_offset": 2048, 00:12:51.206 "data_size": 63488 00:12:51.206 }, 00:12:51.206 { 00:12:51.206 "name": "BaseBdev4", 00:12:51.206 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:51.206 "is_configured": true, 00:12:51.206 "data_offset": 2048, 00:12:51.206 "data_size": 63488 00:12:51.206 } 00:12:51.206 ] 00:12:51.206 }' 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.206 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:51.206 "name": "raid_bdev1", 00:12:51.206 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:51.206 "strip_size_kb": 0, 00:12:51.206 "state": "online", 00:12:51.206 "raid_level": "raid1", 00:12:51.206 "superblock": true, 00:12:51.206 "num_base_bdevs": 4, 00:12:51.206 "num_base_bdevs_discovered": 3, 00:12:51.206 "num_base_bdevs_operational": 3, 00:12:51.206 "base_bdevs_list": [ 00:12:51.206 { 00:12:51.206 "name": "spare", 00:12:51.206 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:51.206 "is_configured": true, 00:12:51.206 "data_offset": 2048, 00:12:51.206 "data_size": 63488 00:12:51.206 }, 00:12:51.206 { 00:12:51.206 "name": null, 00:12:51.206 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:51.206 "is_configured": false, 00:12:51.207 "data_offset": 0, 00:12:51.207 "data_size": 63488 00:12:51.207 }, 00:12:51.207 { 00:12:51.207 "name": "BaseBdev3", 00:12:51.207 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:51.207 "is_configured": true, 00:12:51.207 "data_offset": 2048, 00:12:51.207 "data_size": 63488 00:12:51.207 }, 00:12:51.207 { 00:12:51.207 "name": "BaseBdev4", 00:12:51.207 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:51.207 "is_configured": true, 00:12:51.207 "data_offset": 2048, 00:12:51.207 "data_size": 63488 00:12:51.207 } 00:12:51.207 ] 00:12:51.207 }' 00:12:51.207 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:51.207 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.778 82.62 IOPS, 247.88 MiB/s [2024-12-07T05:40:25.146Z] 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:51.778 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.778 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.778 [2024-12-07 05:40:24.936486] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:51.778 [2024-12-07 05:40:24.936556] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:51.778 00:12:51.778 Latency(us) 00:12:51.778 [2024-12-07T05:40:25.146Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:51.778 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:51.778 raid_bdev1 : 8.06 82.28 246.83 0.00 0.00 15845.83 282.61 112641.79 00:12:51.778 [2024-12-07T05:40:25.146Z] =================================================================================================================== 00:12:51.778 [2024-12-07T05:40:25.146Z] Total : 82.28 246.83 0.00 0.00 15845.83 282.61 112641.79 00:12:51.778 { 00:12:51.778 "results": [ 00:12:51.778 { 00:12:51.778 "job": "raid_bdev1", 00:12:51.778 "core_mask": "0x1", 00:12:51.778 "workload": "randrw", 00:12:51.778 "percentage": 50, 00:12:51.778 "status": "finished", 00:12:51.778 "queue_depth": 2, 00:12:51.778 "io_size": 3145728, 00:12:51.778 "runtime": 8.058299, 00:12:51.778 "iops": 82.27542810213421, 00:12:51.778 "mibps": 246.82628430640264, 00:12:51.778 "io_failed": 0, 00:12:51.778 "io_timeout": 0, 00:12:51.778 "avg_latency_us": 15845.826547320306, 00:12:51.778 "min_latency_us": 282.6061135371179, 00:12:51.778 "max_latency_us": 112641.78864628822 00:12:51.778 } 00:12:51.778 ], 00:12:51.778 "core_count": 1 00:12:51.778 } 00:12:51.778 [2024-12-07 05:40:24.951701] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:51.778 [2024-12-07 05:40:24.951759] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:51.778 [2024-12-07 05:40:24.951855] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:51.778 [2024-12-07 05:40:24.951865] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:51.778 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.778 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.778 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:51.778 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.778 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.778 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.778 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:51.778 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:51.778 05:40:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:51.778 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:51.778 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:51.778 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:51.778 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:51.778 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:51.778 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:51.778 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:51.778 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:51.778 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:51.778 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:52.038 /dev/nbd0 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:52.038 1+0 records in 00:12:52.038 1+0 records out 00:12:52.038 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000415147 s, 9.9 MB/s 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:52.038 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:12:52.297 /dev/nbd1 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:52.297 1+0 records in 00:12:52.297 1+0 records out 00:12:52.297 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000511886 s, 8.0 MB/s 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:52.297 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:52.298 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:52.557 05:40:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:12:52.817 /dev/nbd1 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:52.817 1+0 records in 00:12:52.817 1+0 records out 00:12:52.817 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000401798 s, 10.2 MB/s 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:52.817 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:53.077 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.337 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.337 [2024-12-07 05:40:26.498443] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:53.337 [2024-12-07 05:40:26.498503] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:53.337 [2024-12-07 05:40:26.498529] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:12:53.337 [2024-12-07 05:40:26.498538] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:53.337 [2024-12-07 05:40:26.500797] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:53.337 spare 00:12:53.337 [2024-12-07 05:40:26.500893] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:53.337 [2024-12-07 05:40:26.500988] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:53.337 [2024-12-07 05:40:26.501028] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:53.337 [2024-12-07 05:40:26.501154] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:53.337 [2024-12-07 05:40:26.501260] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.338 [2024-12-07 05:40:26.601151] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:12:53.338 [2024-12-07 05:40:26.601177] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:53.338 [2024-12-07 05:40:26.601446] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000337b0 00:12:53.338 [2024-12-07 05:40:26.601598] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:12:53.338 [2024-12-07 05:40:26.601618] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:12:53.338 [2024-12-07 05:40:26.601787] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:53.338 "name": "raid_bdev1", 00:12:53.338 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:53.338 "strip_size_kb": 0, 00:12:53.338 "state": "online", 00:12:53.338 "raid_level": "raid1", 00:12:53.338 "superblock": true, 00:12:53.338 "num_base_bdevs": 4, 00:12:53.338 "num_base_bdevs_discovered": 3, 00:12:53.338 "num_base_bdevs_operational": 3, 00:12:53.338 "base_bdevs_list": [ 00:12:53.338 { 00:12:53.338 "name": "spare", 00:12:53.338 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:53.338 "is_configured": true, 00:12:53.338 "data_offset": 2048, 00:12:53.338 "data_size": 63488 00:12:53.338 }, 00:12:53.338 { 00:12:53.338 "name": null, 00:12:53.338 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:53.338 "is_configured": false, 00:12:53.338 "data_offset": 2048, 00:12:53.338 "data_size": 63488 00:12:53.338 }, 00:12:53.338 { 00:12:53.338 "name": "BaseBdev3", 00:12:53.338 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:53.338 "is_configured": true, 00:12:53.338 "data_offset": 2048, 00:12:53.338 "data_size": 63488 00:12:53.338 }, 00:12:53.338 { 00:12:53.338 "name": "BaseBdev4", 00:12:53.338 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:53.338 "is_configured": true, 00:12:53.338 "data_offset": 2048, 00:12:53.338 "data_size": 63488 00:12:53.338 } 00:12:53.338 ] 00:12:53.338 }' 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:53.338 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.598 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:53.598 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:53.598 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:53.598 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:53.598 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:53.598 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.598 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.598 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.598 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.859 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.859 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:53.859 "name": "raid_bdev1", 00:12:53.859 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:53.859 "strip_size_kb": 0, 00:12:53.859 "state": "online", 00:12:53.859 "raid_level": "raid1", 00:12:53.859 "superblock": true, 00:12:53.859 "num_base_bdevs": 4, 00:12:53.859 "num_base_bdevs_discovered": 3, 00:12:53.859 "num_base_bdevs_operational": 3, 00:12:53.859 "base_bdevs_list": [ 00:12:53.859 { 00:12:53.859 "name": "spare", 00:12:53.859 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:53.859 "is_configured": true, 00:12:53.859 "data_offset": 2048, 00:12:53.859 "data_size": 63488 00:12:53.859 }, 00:12:53.859 { 00:12:53.859 "name": null, 00:12:53.859 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:53.859 "is_configured": false, 00:12:53.859 "data_offset": 2048, 00:12:53.859 "data_size": 63488 00:12:53.859 }, 00:12:53.859 { 00:12:53.859 "name": "BaseBdev3", 00:12:53.859 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:53.859 "is_configured": true, 00:12:53.859 "data_offset": 2048, 00:12:53.859 "data_size": 63488 00:12:53.859 }, 00:12:53.859 { 00:12:53.859 "name": "BaseBdev4", 00:12:53.859 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:53.859 "is_configured": true, 00:12:53.859 "data_offset": 2048, 00:12:53.859 "data_size": 63488 00:12:53.859 } 00:12:53.859 ] 00:12:53.859 }' 00:12:53.859 05:40:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.859 [2024-12-07 05:40:27.093572] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.859 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.860 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.860 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:53.860 "name": "raid_bdev1", 00:12:53.860 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:53.860 "strip_size_kb": 0, 00:12:53.860 "state": "online", 00:12:53.860 "raid_level": "raid1", 00:12:53.860 "superblock": true, 00:12:53.860 "num_base_bdevs": 4, 00:12:53.860 "num_base_bdevs_discovered": 2, 00:12:53.860 "num_base_bdevs_operational": 2, 00:12:53.860 "base_bdevs_list": [ 00:12:53.860 { 00:12:53.860 "name": null, 00:12:53.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:53.860 "is_configured": false, 00:12:53.860 "data_offset": 0, 00:12:53.860 "data_size": 63488 00:12:53.860 }, 00:12:53.860 { 00:12:53.860 "name": null, 00:12:53.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:53.860 "is_configured": false, 00:12:53.860 "data_offset": 2048, 00:12:53.860 "data_size": 63488 00:12:53.860 }, 00:12:53.860 { 00:12:53.860 "name": "BaseBdev3", 00:12:53.860 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:53.860 "is_configured": true, 00:12:53.860 "data_offset": 2048, 00:12:53.860 "data_size": 63488 00:12:53.860 }, 00:12:53.860 { 00:12:53.860 "name": "BaseBdev4", 00:12:53.860 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:53.860 "is_configured": true, 00:12:53.860 "data_offset": 2048, 00:12:53.860 "data_size": 63488 00:12:53.860 } 00:12:53.860 ] 00:12:53.860 }' 00:12:53.860 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:53.860 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:54.430 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:54.430 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.430 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:54.431 [2024-12-07 05:40:27.500932] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:54.431 [2024-12-07 05:40:27.501207] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:54.431 [2024-12-07 05:40:27.501236] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:54.431 [2024-12-07 05:40:27.501290] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:54.431 [2024-12-07 05:40:27.505696] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033880 00:12:54.431 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.431 05:40:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:54.431 [2024-12-07 05:40:27.507707] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:55.373 "name": "raid_bdev1", 00:12:55.373 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:55.373 "strip_size_kb": 0, 00:12:55.373 "state": "online", 00:12:55.373 "raid_level": "raid1", 00:12:55.373 "superblock": true, 00:12:55.373 "num_base_bdevs": 4, 00:12:55.373 "num_base_bdevs_discovered": 3, 00:12:55.373 "num_base_bdevs_operational": 3, 00:12:55.373 "process": { 00:12:55.373 "type": "rebuild", 00:12:55.373 "target": "spare", 00:12:55.373 "progress": { 00:12:55.373 "blocks": 20480, 00:12:55.373 "percent": 32 00:12:55.373 } 00:12:55.373 }, 00:12:55.373 "base_bdevs_list": [ 00:12:55.373 { 00:12:55.373 "name": "spare", 00:12:55.373 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:55.373 "is_configured": true, 00:12:55.373 "data_offset": 2048, 00:12:55.373 "data_size": 63488 00:12:55.373 }, 00:12:55.373 { 00:12:55.373 "name": null, 00:12:55.373 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.373 "is_configured": false, 00:12:55.373 "data_offset": 2048, 00:12:55.373 "data_size": 63488 00:12:55.373 }, 00:12:55.373 { 00:12:55.373 "name": "BaseBdev3", 00:12:55.373 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:55.373 "is_configured": true, 00:12:55.373 "data_offset": 2048, 00:12:55.373 "data_size": 63488 00:12:55.373 }, 00:12:55.373 { 00:12:55.373 "name": "BaseBdev4", 00:12:55.373 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:55.373 "is_configured": true, 00:12:55.373 "data_offset": 2048, 00:12:55.373 "data_size": 63488 00:12:55.373 } 00:12:55.373 ] 00:12:55.373 }' 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.373 [2024-12-07 05:40:28.644163] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:55.373 [2024-12-07 05:40:28.712484] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:55.373 [2024-12-07 05:40:28.712583] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:55.373 [2024-12-07 05:40:28.712620] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:55.373 [2024-12-07 05:40:28.712629] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.373 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.633 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.633 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:55.633 "name": "raid_bdev1", 00:12:55.633 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:55.633 "strip_size_kb": 0, 00:12:55.633 "state": "online", 00:12:55.633 "raid_level": "raid1", 00:12:55.633 "superblock": true, 00:12:55.633 "num_base_bdevs": 4, 00:12:55.633 "num_base_bdevs_discovered": 2, 00:12:55.633 "num_base_bdevs_operational": 2, 00:12:55.633 "base_bdevs_list": [ 00:12:55.633 { 00:12:55.633 "name": null, 00:12:55.633 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.633 "is_configured": false, 00:12:55.633 "data_offset": 0, 00:12:55.633 "data_size": 63488 00:12:55.633 }, 00:12:55.633 { 00:12:55.633 "name": null, 00:12:55.633 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.633 "is_configured": false, 00:12:55.633 "data_offset": 2048, 00:12:55.633 "data_size": 63488 00:12:55.633 }, 00:12:55.633 { 00:12:55.633 "name": "BaseBdev3", 00:12:55.633 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:55.633 "is_configured": true, 00:12:55.633 "data_offset": 2048, 00:12:55.633 "data_size": 63488 00:12:55.633 }, 00:12:55.633 { 00:12:55.633 "name": "BaseBdev4", 00:12:55.633 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:55.633 "is_configured": true, 00:12:55.633 "data_offset": 2048, 00:12:55.633 "data_size": 63488 00:12:55.633 } 00:12:55.633 ] 00:12:55.633 }' 00:12:55.633 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:55.633 05:40:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.893 05:40:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:55.893 05:40:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.893 05:40:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.893 [2024-12-07 05:40:29.148480] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:55.893 [2024-12-07 05:40:29.148588] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:55.893 [2024-12-07 05:40:29.148642] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:12:55.893 [2024-12-07 05:40:29.148693] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:55.893 [2024-12-07 05:40:29.149146] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:55.893 [2024-12-07 05:40:29.149205] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:55.893 [2024-12-07 05:40:29.149331] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:55.893 [2024-12-07 05:40:29.149371] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:55.893 [2024-12-07 05:40:29.149414] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:55.893 [2024-12-07 05:40:29.149472] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:55.893 [2024-12-07 05:40:29.153603] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033950 00:12:55.893 spare 00:12:55.893 05:40:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.893 05:40:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:55.893 [2024-12-07 05:40:29.155592] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:56.833 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:56.833 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:56.833 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:56.833 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:56.833 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:56.833 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.833 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.833 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:56.833 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:56.833 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:57.094 "name": "raid_bdev1", 00:12:57.094 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:57.094 "strip_size_kb": 0, 00:12:57.094 "state": "online", 00:12:57.094 "raid_level": "raid1", 00:12:57.094 "superblock": true, 00:12:57.094 "num_base_bdevs": 4, 00:12:57.094 "num_base_bdevs_discovered": 3, 00:12:57.094 "num_base_bdevs_operational": 3, 00:12:57.094 "process": { 00:12:57.094 "type": "rebuild", 00:12:57.094 "target": "spare", 00:12:57.094 "progress": { 00:12:57.094 "blocks": 20480, 00:12:57.094 "percent": 32 00:12:57.094 } 00:12:57.094 }, 00:12:57.094 "base_bdevs_list": [ 00:12:57.094 { 00:12:57.094 "name": "spare", 00:12:57.094 "uuid": "39dcd5bf-9282-53c5-ab70-e7d1be21a1e2", 00:12:57.094 "is_configured": true, 00:12:57.094 "data_offset": 2048, 00:12:57.094 "data_size": 63488 00:12:57.094 }, 00:12:57.094 { 00:12:57.094 "name": null, 00:12:57.094 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.094 "is_configured": false, 00:12:57.094 "data_offset": 2048, 00:12:57.094 "data_size": 63488 00:12:57.094 }, 00:12:57.094 { 00:12:57.094 "name": "BaseBdev3", 00:12:57.094 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:57.094 "is_configured": true, 00:12:57.094 "data_offset": 2048, 00:12:57.094 "data_size": 63488 00:12:57.094 }, 00:12:57.094 { 00:12:57.094 "name": "BaseBdev4", 00:12:57.094 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:57.094 "is_configured": true, 00:12:57.094 "data_offset": 2048, 00:12:57.094 "data_size": 63488 00:12:57.094 } 00:12:57.094 ] 00:12:57.094 }' 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:57.094 [2024-12-07 05:40:30.319911] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:57.094 [2024-12-07 05:40:30.359738] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:57.094 [2024-12-07 05:40:30.359814] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:57.094 [2024-12-07 05:40:30.359830] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:57.094 [2024-12-07 05:40:30.359839] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:57.094 "name": "raid_bdev1", 00:12:57.094 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:57.094 "strip_size_kb": 0, 00:12:57.094 "state": "online", 00:12:57.094 "raid_level": "raid1", 00:12:57.094 "superblock": true, 00:12:57.094 "num_base_bdevs": 4, 00:12:57.094 "num_base_bdevs_discovered": 2, 00:12:57.094 "num_base_bdevs_operational": 2, 00:12:57.094 "base_bdevs_list": [ 00:12:57.094 { 00:12:57.094 "name": null, 00:12:57.094 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.094 "is_configured": false, 00:12:57.094 "data_offset": 0, 00:12:57.094 "data_size": 63488 00:12:57.094 }, 00:12:57.094 { 00:12:57.094 "name": null, 00:12:57.094 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.094 "is_configured": false, 00:12:57.094 "data_offset": 2048, 00:12:57.094 "data_size": 63488 00:12:57.094 }, 00:12:57.094 { 00:12:57.094 "name": "BaseBdev3", 00:12:57.094 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:57.094 "is_configured": true, 00:12:57.094 "data_offset": 2048, 00:12:57.094 "data_size": 63488 00:12:57.094 }, 00:12:57.094 { 00:12:57.094 "name": "BaseBdev4", 00:12:57.094 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:57.094 "is_configured": true, 00:12:57.094 "data_offset": 2048, 00:12:57.094 "data_size": 63488 00:12:57.094 } 00:12:57.094 ] 00:12:57.094 }' 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:57.094 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:57.666 "name": "raid_bdev1", 00:12:57.666 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:57.666 "strip_size_kb": 0, 00:12:57.666 "state": "online", 00:12:57.666 "raid_level": "raid1", 00:12:57.666 "superblock": true, 00:12:57.666 "num_base_bdevs": 4, 00:12:57.666 "num_base_bdevs_discovered": 2, 00:12:57.666 "num_base_bdevs_operational": 2, 00:12:57.666 "base_bdevs_list": [ 00:12:57.666 { 00:12:57.666 "name": null, 00:12:57.666 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.666 "is_configured": false, 00:12:57.666 "data_offset": 0, 00:12:57.666 "data_size": 63488 00:12:57.666 }, 00:12:57.666 { 00:12:57.666 "name": null, 00:12:57.666 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.666 "is_configured": false, 00:12:57.666 "data_offset": 2048, 00:12:57.666 "data_size": 63488 00:12:57.666 }, 00:12:57.666 { 00:12:57.666 "name": "BaseBdev3", 00:12:57.666 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:57.666 "is_configured": true, 00:12:57.666 "data_offset": 2048, 00:12:57.666 "data_size": 63488 00:12:57.666 }, 00:12:57.666 { 00:12:57.666 "name": "BaseBdev4", 00:12:57.666 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:57.666 "is_configured": true, 00:12:57.666 "data_offset": 2048, 00:12:57.666 "data_size": 63488 00:12:57.666 } 00:12:57.666 ] 00:12:57.666 }' 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:57.666 [2024-12-07 05:40:30.947347] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:57.666 [2024-12-07 05:40:30.947458] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:57.666 [2024-12-07 05:40:30.947484] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:12:57.666 [2024-12-07 05:40:30.947494] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:57.666 [2024-12-07 05:40:30.947901] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:57.666 [2024-12-07 05:40:30.947926] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:57.666 [2024-12-07 05:40:30.947996] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:57.666 [2024-12-07 05:40:30.948011] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:57.666 [2024-12-07 05:40:30.948021] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:57.666 [2024-12-07 05:40:30.948034] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:57.666 BaseBdev1 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.666 05:40:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.606 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.867 05:40:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.867 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:58.867 "name": "raid_bdev1", 00:12:58.867 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:58.867 "strip_size_kb": 0, 00:12:58.867 "state": "online", 00:12:58.867 "raid_level": "raid1", 00:12:58.867 "superblock": true, 00:12:58.867 "num_base_bdevs": 4, 00:12:58.867 "num_base_bdevs_discovered": 2, 00:12:58.867 "num_base_bdevs_operational": 2, 00:12:58.867 "base_bdevs_list": [ 00:12:58.867 { 00:12:58.867 "name": null, 00:12:58.867 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:58.867 "is_configured": false, 00:12:58.867 "data_offset": 0, 00:12:58.867 "data_size": 63488 00:12:58.867 }, 00:12:58.867 { 00:12:58.867 "name": null, 00:12:58.867 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:58.867 "is_configured": false, 00:12:58.867 "data_offset": 2048, 00:12:58.867 "data_size": 63488 00:12:58.867 }, 00:12:58.867 { 00:12:58.867 "name": "BaseBdev3", 00:12:58.867 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:58.867 "is_configured": true, 00:12:58.867 "data_offset": 2048, 00:12:58.867 "data_size": 63488 00:12:58.867 }, 00:12:58.867 { 00:12:58.867 "name": "BaseBdev4", 00:12:58.867 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:58.867 "is_configured": true, 00:12:58.867 "data_offset": 2048, 00:12:58.867 "data_size": 63488 00:12:58.867 } 00:12:58.867 ] 00:12:58.867 }' 00:12:58.867 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:58.867 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.128 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:59.128 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:59.128 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:59.128 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:59.128 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:59.128 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.128 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.128 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.128 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.128 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.128 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:59.128 "name": "raid_bdev1", 00:12:59.128 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:12:59.128 "strip_size_kb": 0, 00:12:59.128 "state": "online", 00:12:59.128 "raid_level": "raid1", 00:12:59.128 "superblock": true, 00:12:59.128 "num_base_bdevs": 4, 00:12:59.128 "num_base_bdevs_discovered": 2, 00:12:59.128 "num_base_bdevs_operational": 2, 00:12:59.128 "base_bdevs_list": [ 00:12:59.128 { 00:12:59.128 "name": null, 00:12:59.128 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.128 "is_configured": false, 00:12:59.128 "data_offset": 0, 00:12:59.129 "data_size": 63488 00:12:59.129 }, 00:12:59.129 { 00:12:59.129 "name": null, 00:12:59.129 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.129 "is_configured": false, 00:12:59.129 "data_offset": 2048, 00:12:59.129 "data_size": 63488 00:12:59.129 }, 00:12:59.129 { 00:12:59.129 "name": "BaseBdev3", 00:12:59.129 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:12:59.129 "is_configured": true, 00:12:59.129 "data_offset": 2048, 00:12:59.129 "data_size": 63488 00:12:59.129 }, 00:12:59.129 { 00:12:59.129 "name": "BaseBdev4", 00:12:59.129 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:12:59.129 "is_configured": true, 00:12:59.129 "data_offset": 2048, 00:12:59.129 "data_size": 63488 00:12:59.129 } 00:12:59.129 ] 00:12:59.129 }' 00:12:59.129 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.389 [2024-12-07 05:40:32.580769] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:59.389 [2024-12-07 05:40:32.580942] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:59.389 [2024-12-07 05:40:32.580956] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:59.389 request: 00:12:59.389 { 00:12:59.389 "base_bdev": "BaseBdev1", 00:12:59.389 "raid_bdev": "raid_bdev1", 00:12:59.389 "method": "bdev_raid_add_base_bdev", 00:12:59.389 "req_id": 1 00:12:59.389 } 00:12:59.389 Got JSON-RPC error response 00:12:59.389 response: 00:12:59.389 { 00:12:59.389 "code": -22, 00:12:59.389 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:59.389 } 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:59.389 05:40:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:00.345 "name": "raid_bdev1", 00:13:00.345 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:13:00.345 "strip_size_kb": 0, 00:13:00.345 "state": "online", 00:13:00.345 "raid_level": "raid1", 00:13:00.345 "superblock": true, 00:13:00.345 "num_base_bdevs": 4, 00:13:00.345 "num_base_bdevs_discovered": 2, 00:13:00.345 "num_base_bdevs_operational": 2, 00:13:00.345 "base_bdevs_list": [ 00:13:00.345 { 00:13:00.345 "name": null, 00:13:00.345 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:00.345 "is_configured": false, 00:13:00.345 "data_offset": 0, 00:13:00.345 "data_size": 63488 00:13:00.345 }, 00:13:00.345 { 00:13:00.345 "name": null, 00:13:00.345 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:00.345 "is_configured": false, 00:13:00.345 "data_offset": 2048, 00:13:00.345 "data_size": 63488 00:13:00.345 }, 00:13:00.345 { 00:13:00.345 "name": "BaseBdev3", 00:13:00.345 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:13:00.345 "is_configured": true, 00:13:00.345 "data_offset": 2048, 00:13:00.345 "data_size": 63488 00:13:00.345 }, 00:13:00.345 { 00:13:00.345 "name": "BaseBdev4", 00:13:00.345 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:13:00.345 "is_configured": true, 00:13:00.345 "data_offset": 2048, 00:13:00.345 "data_size": 63488 00:13:00.345 } 00:13:00.345 ] 00:13:00.345 }' 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:00.345 05:40:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:00.915 "name": "raid_bdev1", 00:13:00.915 "uuid": "27a93d37-f2f2-46d7-a402-a509c0d653f0", 00:13:00.915 "strip_size_kb": 0, 00:13:00.915 "state": "online", 00:13:00.915 "raid_level": "raid1", 00:13:00.915 "superblock": true, 00:13:00.915 "num_base_bdevs": 4, 00:13:00.915 "num_base_bdevs_discovered": 2, 00:13:00.915 "num_base_bdevs_operational": 2, 00:13:00.915 "base_bdevs_list": [ 00:13:00.915 { 00:13:00.915 "name": null, 00:13:00.915 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:00.915 "is_configured": false, 00:13:00.915 "data_offset": 0, 00:13:00.915 "data_size": 63488 00:13:00.915 }, 00:13:00.915 { 00:13:00.915 "name": null, 00:13:00.915 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:00.915 "is_configured": false, 00:13:00.915 "data_offset": 2048, 00:13:00.915 "data_size": 63488 00:13:00.915 }, 00:13:00.915 { 00:13:00.915 "name": "BaseBdev3", 00:13:00.915 "uuid": "f8ecc51c-6a15-57aa-be8c-37bbdfb9144b", 00:13:00.915 "is_configured": true, 00:13:00.915 "data_offset": 2048, 00:13:00.915 "data_size": 63488 00:13:00.915 }, 00:13:00.915 { 00:13:00.915 "name": "BaseBdev4", 00:13:00.915 "uuid": "7eaff4d9-428b-5dd1-b969-55e727b4d21b", 00:13:00.915 "is_configured": true, 00:13:00.915 "data_offset": 2048, 00:13:00.915 "data_size": 63488 00:13:00.915 } 00:13:00.915 ] 00:13:00.915 }' 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:00.915 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:00.916 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 89366 00:13:00.916 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 89366 ']' 00:13:00.916 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 89366 00:13:00.916 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:13:00.916 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:00.916 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 89366 00:13:00.916 killing process with pid 89366 00:13:00.916 Received shutdown signal, test time was about 17.297625 seconds 00:13:00.916 00:13:00.916 Latency(us) 00:13:00.916 [2024-12-07T05:40:34.284Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:00.916 [2024-12-07T05:40:34.284Z] =================================================================================================================== 00:13:00.916 [2024-12-07T05:40:34.284Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:00.916 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:00.916 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:00.916 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 89366' 00:13:00.916 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 89366 00:13:00.916 [2024-12-07 05:40:34.169919] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:00.916 [2024-12-07 05:40:34.170043] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:00.916 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 89366 00:13:00.916 [2024-12-07 05:40:34.170114] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:00.916 [2024-12-07 05:40:34.170136] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:13:00.916 [2024-12-07 05:40:34.216691] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:01.176 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:01.176 00:13:01.176 real 0m19.238s 00:13:01.176 user 0m25.579s 00:13:01.176 sys 0m2.300s 00:13:01.176 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:01.176 05:40:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.176 ************************************ 00:13:01.176 END TEST raid_rebuild_test_sb_io 00:13:01.176 ************************************ 00:13:01.176 05:40:34 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:13:01.176 05:40:34 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:13:01.176 05:40:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:13:01.176 05:40:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:01.176 05:40:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:01.176 ************************************ 00:13:01.176 START TEST raid5f_state_function_test 00:13:01.176 ************************************ 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 false 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=90071 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90071' 00:13:01.176 Process raid pid: 90071 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 90071 00:13:01.176 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 90071 ']' 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:01.176 05:40:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:01.437 [2024-12-07 05:40:34.591250] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:13:01.437 [2024-12-07 05:40:34.591452] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:01.437 [2024-12-07 05:40:34.744981] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:01.437 [2024-12-07 05:40:34.769399] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:01.701 [2024-12-07 05:40:34.811405] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:01.701 [2024-12-07 05:40:34.811441] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.274 [2024-12-07 05:40:35.413888] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:02.274 [2024-12-07 05:40:35.414011] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:02.274 [2024-12-07 05:40:35.414026] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:02.274 [2024-12-07 05:40:35.414035] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:02.274 [2024-12-07 05:40:35.414042] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:02.274 [2024-12-07 05:40:35.414054] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.274 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.275 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:02.275 "name": "Existed_Raid", 00:13:02.275 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.275 "strip_size_kb": 64, 00:13:02.275 "state": "configuring", 00:13:02.275 "raid_level": "raid5f", 00:13:02.275 "superblock": false, 00:13:02.275 "num_base_bdevs": 3, 00:13:02.275 "num_base_bdevs_discovered": 0, 00:13:02.275 "num_base_bdevs_operational": 3, 00:13:02.275 "base_bdevs_list": [ 00:13:02.275 { 00:13:02.275 "name": "BaseBdev1", 00:13:02.275 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.275 "is_configured": false, 00:13:02.275 "data_offset": 0, 00:13:02.275 "data_size": 0 00:13:02.275 }, 00:13:02.275 { 00:13:02.275 "name": "BaseBdev2", 00:13:02.275 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.275 "is_configured": false, 00:13:02.275 "data_offset": 0, 00:13:02.275 "data_size": 0 00:13:02.275 }, 00:13:02.275 { 00:13:02.275 "name": "BaseBdev3", 00:13:02.275 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.275 "is_configured": false, 00:13:02.275 "data_offset": 0, 00:13:02.275 "data_size": 0 00:13:02.275 } 00:13:02.275 ] 00:13:02.275 }' 00:13:02.275 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:02.275 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.533 [2024-12-07 05:40:35.825080] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:02.533 [2024-12-07 05:40:35.825161] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.533 [2024-12-07 05:40:35.837087] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:02.533 [2024-12-07 05:40:35.837170] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:02.533 [2024-12-07 05:40:35.837197] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:02.533 [2024-12-07 05:40:35.837219] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:02.533 [2024-12-07 05:40:35.837237] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:02.533 [2024-12-07 05:40:35.837258] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.533 [2024-12-07 05:40:35.857861] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:02.533 BaseBdev1 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.533 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.533 [ 00:13:02.533 { 00:13:02.533 "name": "BaseBdev1", 00:13:02.533 "aliases": [ 00:13:02.533 "e071571f-9a6a-4f6b-bc34-ea32ef9f6f7a" 00:13:02.533 ], 00:13:02.533 "product_name": "Malloc disk", 00:13:02.533 "block_size": 512, 00:13:02.533 "num_blocks": 65536, 00:13:02.533 "uuid": "e071571f-9a6a-4f6b-bc34-ea32ef9f6f7a", 00:13:02.533 "assigned_rate_limits": { 00:13:02.533 "rw_ios_per_sec": 0, 00:13:02.533 "rw_mbytes_per_sec": 0, 00:13:02.533 "r_mbytes_per_sec": 0, 00:13:02.533 "w_mbytes_per_sec": 0 00:13:02.533 }, 00:13:02.533 "claimed": true, 00:13:02.533 "claim_type": "exclusive_write", 00:13:02.533 "zoned": false, 00:13:02.533 "supported_io_types": { 00:13:02.533 "read": true, 00:13:02.534 "write": true, 00:13:02.534 "unmap": true, 00:13:02.534 "flush": true, 00:13:02.534 "reset": true, 00:13:02.534 "nvme_admin": false, 00:13:02.534 "nvme_io": false, 00:13:02.534 "nvme_io_md": false, 00:13:02.534 "write_zeroes": true, 00:13:02.534 "zcopy": true, 00:13:02.534 "get_zone_info": false, 00:13:02.534 "zone_management": false, 00:13:02.534 "zone_append": false, 00:13:02.534 "compare": false, 00:13:02.534 "compare_and_write": false, 00:13:02.534 "abort": true, 00:13:02.534 "seek_hole": false, 00:13:02.534 "seek_data": false, 00:13:02.534 "copy": true, 00:13:02.534 "nvme_iov_md": false 00:13:02.534 }, 00:13:02.534 "memory_domains": [ 00:13:02.534 { 00:13:02.534 "dma_device_id": "system", 00:13:02.534 "dma_device_type": 1 00:13:02.534 }, 00:13:02.534 { 00:13:02.534 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:02.534 "dma_device_type": 2 00:13:02.792 } 00:13:02.792 ], 00:13:02.792 "driver_specific": {} 00:13:02.792 } 00:13:02.792 ] 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:02.792 "name": "Existed_Raid", 00:13:02.792 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.792 "strip_size_kb": 64, 00:13:02.792 "state": "configuring", 00:13:02.792 "raid_level": "raid5f", 00:13:02.792 "superblock": false, 00:13:02.792 "num_base_bdevs": 3, 00:13:02.792 "num_base_bdevs_discovered": 1, 00:13:02.792 "num_base_bdevs_operational": 3, 00:13:02.792 "base_bdevs_list": [ 00:13:02.792 { 00:13:02.792 "name": "BaseBdev1", 00:13:02.792 "uuid": "e071571f-9a6a-4f6b-bc34-ea32ef9f6f7a", 00:13:02.792 "is_configured": true, 00:13:02.792 "data_offset": 0, 00:13:02.792 "data_size": 65536 00:13:02.792 }, 00:13:02.792 { 00:13:02.792 "name": "BaseBdev2", 00:13:02.792 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.792 "is_configured": false, 00:13:02.792 "data_offset": 0, 00:13:02.792 "data_size": 0 00:13:02.792 }, 00:13:02.792 { 00:13:02.792 "name": "BaseBdev3", 00:13:02.792 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.792 "is_configured": false, 00:13:02.792 "data_offset": 0, 00:13:02.792 "data_size": 0 00:13:02.792 } 00:13:02.792 ] 00:13:02.792 }' 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:02.792 05:40:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.051 [2024-12-07 05:40:36.313132] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:03.051 [2024-12-07 05:40:36.313183] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.051 [2024-12-07 05:40:36.325148] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:03.051 [2024-12-07 05:40:36.327041] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:03.051 [2024-12-07 05:40:36.327113] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:03.051 [2024-12-07 05:40:36.327148] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:03.051 [2024-12-07 05:40:36.327174] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:03.051 "name": "Existed_Raid", 00:13:03.051 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.051 "strip_size_kb": 64, 00:13:03.051 "state": "configuring", 00:13:03.051 "raid_level": "raid5f", 00:13:03.051 "superblock": false, 00:13:03.051 "num_base_bdevs": 3, 00:13:03.051 "num_base_bdevs_discovered": 1, 00:13:03.051 "num_base_bdevs_operational": 3, 00:13:03.051 "base_bdevs_list": [ 00:13:03.051 { 00:13:03.051 "name": "BaseBdev1", 00:13:03.051 "uuid": "e071571f-9a6a-4f6b-bc34-ea32ef9f6f7a", 00:13:03.051 "is_configured": true, 00:13:03.051 "data_offset": 0, 00:13:03.051 "data_size": 65536 00:13:03.051 }, 00:13:03.051 { 00:13:03.051 "name": "BaseBdev2", 00:13:03.051 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.051 "is_configured": false, 00:13:03.051 "data_offset": 0, 00:13:03.051 "data_size": 0 00:13:03.051 }, 00:13:03.051 { 00:13:03.051 "name": "BaseBdev3", 00:13:03.051 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.051 "is_configured": false, 00:13:03.051 "data_offset": 0, 00:13:03.051 "data_size": 0 00:13:03.051 } 00:13:03.051 ] 00:13:03.051 }' 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:03.051 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.621 [2024-12-07 05:40:36.739451] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:03.621 BaseBdev2 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.621 [ 00:13:03.621 { 00:13:03.621 "name": "BaseBdev2", 00:13:03.621 "aliases": [ 00:13:03.621 "6c900b77-85d4-46ba-85c9-2278c3658f8e" 00:13:03.621 ], 00:13:03.621 "product_name": "Malloc disk", 00:13:03.621 "block_size": 512, 00:13:03.621 "num_blocks": 65536, 00:13:03.621 "uuid": "6c900b77-85d4-46ba-85c9-2278c3658f8e", 00:13:03.621 "assigned_rate_limits": { 00:13:03.621 "rw_ios_per_sec": 0, 00:13:03.621 "rw_mbytes_per_sec": 0, 00:13:03.621 "r_mbytes_per_sec": 0, 00:13:03.621 "w_mbytes_per_sec": 0 00:13:03.621 }, 00:13:03.621 "claimed": true, 00:13:03.621 "claim_type": "exclusive_write", 00:13:03.621 "zoned": false, 00:13:03.621 "supported_io_types": { 00:13:03.621 "read": true, 00:13:03.621 "write": true, 00:13:03.621 "unmap": true, 00:13:03.621 "flush": true, 00:13:03.621 "reset": true, 00:13:03.621 "nvme_admin": false, 00:13:03.621 "nvme_io": false, 00:13:03.621 "nvme_io_md": false, 00:13:03.621 "write_zeroes": true, 00:13:03.621 "zcopy": true, 00:13:03.621 "get_zone_info": false, 00:13:03.621 "zone_management": false, 00:13:03.621 "zone_append": false, 00:13:03.621 "compare": false, 00:13:03.621 "compare_and_write": false, 00:13:03.621 "abort": true, 00:13:03.621 "seek_hole": false, 00:13:03.621 "seek_data": false, 00:13:03.621 "copy": true, 00:13:03.621 "nvme_iov_md": false 00:13:03.621 }, 00:13:03.621 "memory_domains": [ 00:13:03.621 { 00:13:03.621 "dma_device_id": "system", 00:13:03.621 "dma_device_type": 1 00:13:03.621 }, 00:13:03.621 { 00:13:03.621 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:03.621 "dma_device_type": 2 00:13:03.621 } 00:13:03.621 ], 00:13:03.621 "driver_specific": {} 00:13:03.621 } 00:13:03.621 ] 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:03.621 "name": "Existed_Raid", 00:13:03.621 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.621 "strip_size_kb": 64, 00:13:03.621 "state": "configuring", 00:13:03.621 "raid_level": "raid5f", 00:13:03.621 "superblock": false, 00:13:03.621 "num_base_bdevs": 3, 00:13:03.621 "num_base_bdevs_discovered": 2, 00:13:03.621 "num_base_bdevs_operational": 3, 00:13:03.621 "base_bdevs_list": [ 00:13:03.621 { 00:13:03.621 "name": "BaseBdev1", 00:13:03.621 "uuid": "e071571f-9a6a-4f6b-bc34-ea32ef9f6f7a", 00:13:03.621 "is_configured": true, 00:13:03.621 "data_offset": 0, 00:13:03.621 "data_size": 65536 00:13:03.621 }, 00:13:03.621 { 00:13:03.621 "name": "BaseBdev2", 00:13:03.621 "uuid": "6c900b77-85d4-46ba-85c9-2278c3658f8e", 00:13:03.621 "is_configured": true, 00:13:03.621 "data_offset": 0, 00:13:03.621 "data_size": 65536 00:13:03.621 }, 00:13:03.621 { 00:13:03.621 "name": "BaseBdev3", 00:13:03.621 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.621 "is_configured": false, 00:13:03.621 "data_offset": 0, 00:13:03.621 "data_size": 0 00:13:03.621 } 00:13:03.621 ] 00:13:03.621 }' 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:03.621 05:40:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.881 [2024-12-07 05:40:37.222051] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:03.881 [2024-12-07 05:40:37.222194] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:03.881 [2024-12-07 05:40:37.222228] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:03.881 [2024-12-07 05:40:37.222576] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:13:03.881 [2024-12-07 05:40:37.223098] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:03.881 [2024-12-07 05:40:37.223122] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:13:03.881 [2024-12-07 05:40:37.223340] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:03.881 BaseBdev3 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.881 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:04.142 [ 00:13:04.142 { 00:13:04.142 "name": "BaseBdev3", 00:13:04.142 "aliases": [ 00:13:04.142 "bb7aae93-c556-464b-b3c1-df18a1e61017" 00:13:04.142 ], 00:13:04.142 "product_name": "Malloc disk", 00:13:04.142 "block_size": 512, 00:13:04.142 "num_blocks": 65536, 00:13:04.142 "uuid": "bb7aae93-c556-464b-b3c1-df18a1e61017", 00:13:04.142 "assigned_rate_limits": { 00:13:04.142 "rw_ios_per_sec": 0, 00:13:04.142 "rw_mbytes_per_sec": 0, 00:13:04.142 "r_mbytes_per_sec": 0, 00:13:04.142 "w_mbytes_per_sec": 0 00:13:04.142 }, 00:13:04.142 "claimed": true, 00:13:04.142 "claim_type": "exclusive_write", 00:13:04.142 "zoned": false, 00:13:04.142 "supported_io_types": { 00:13:04.142 "read": true, 00:13:04.142 "write": true, 00:13:04.142 "unmap": true, 00:13:04.142 "flush": true, 00:13:04.142 "reset": true, 00:13:04.142 "nvme_admin": false, 00:13:04.142 "nvme_io": false, 00:13:04.142 "nvme_io_md": false, 00:13:04.142 "write_zeroes": true, 00:13:04.142 "zcopy": true, 00:13:04.142 "get_zone_info": false, 00:13:04.142 "zone_management": false, 00:13:04.142 "zone_append": false, 00:13:04.142 "compare": false, 00:13:04.142 "compare_and_write": false, 00:13:04.142 "abort": true, 00:13:04.142 "seek_hole": false, 00:13:04.142 "seek_data": false, 00:13:04.142 "copy": true, 00:13:04.142 "nvme_iov_md": false 00:13:04.142 }, 00:13:04.142 "memory_domains": [ 00:13:04.142 { 00:13:04.142 "dma_device_id": "system", 00:13:04.142 "dma_device_type": 1 00:13:04.142 }, 00:13:04.142 { 00:13:04.142 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:04.142 "dma_device_type": 2 00:13:04.142 } 00:13:04.142 ], 00:13:04.142 "driver_specific": {} 00:13:04.142 } 00:13:04.142 ] 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:04.142 "name": "Existed_Raid", 00:13:04.142 "uuid": "ab7252b3-69b5-4071-a15f-e1e096f77fef", 00:13:04.142 "strip_size_kb": 64, 00:13:04.142 "state": "online", 00:13:04.142 "raid_level": "raid5f", 00:13:04.142 "superblock": false, 00:13:04.142 "num_base_bdevs": 3, 00:13:04.142 "num_base_bdevs_discovered": 3, 00:13:04.142 "num_base_bdevs_operational": 3, 00:13:04.142 "base_bdevs_list": [ 00:13:04.142 { 00:13:04.142 "name": "BaseBdev1", 00:13:04.142 "uuid": "e071571f-9a6a-4f6b-bc34-ea32ef9f6f7a", 00:13:04.142 "is_configured": true, 00:13:04.142 "data_offset": 0, 00:13:04.142 "data_size": 65536 00:13:04.142 }, 00:13:04.142 { 00:13:04.142 "name": "BaseBdev2", 00:13:04.142 "uuid": "6c900b77-85d4-46ba-85c9-2278c3658f8e", 00:13:04.142 "is_configured": true, 00:13:04.142 "data_offset": 0, 00:13:04.142 "data_size": 65536 00:13:04.142 }, 00:13:04.142 { 00:13:04.142 "name": "BaseBdev3", 00:13:04.142 "uuid": "bb7aae93-c556-464b-b3c1-df18a1e61017", 00:13:04.142 "is_configured": true, 00:13:04.142 "data_offset": 0, 00:13:04.142 "data_size": 65536 00:13:04.142 } 00:13:04.142 ] 00:13:04.142 }' 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:04.142 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:04.402 [2024-12-07 05:40:37.669521] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:04.402 "name": "Existed_Raid", 00:13:04.402 "aliases": [ 00:13:04.402 "ab7252b3-69b5-4071-a15f-e1e096f77fef" 00:13:04.402 ], 00:13:04.402 "product_name": "Raid Volume", 00:13:04.402 "block_size": 512, 00:13:04.402 "num_blocks": 131072, 00:13:04.402 "uuid": "ab7252b3-69b5-4071-a15f-e1e096f77fef", 00:13:04.402 "assigned_rate_limits": { 00:13:04.402 "rw_ios_per_sec": 0, 00:13:04.402 "rw_mbytes_per_sec": 0, 00:13:04.402 "r_mbytes_per_sec": 0, 00:13:04.402 "w_mbytes_per_sec": 0 00:13:04.402 }, 00:13:04.402 "claimed": false, 00:13:04.402 "zoned": false, 00:13:04.402 "supported_io_types": { 00:13:04.402 "read": true, 00:13:04.402 "write": true, 00:13:04.402 "unmap": false, 00:13:04.402 "flush": false, 00:13:04.402 "reset": true, 00:13:04.402 "nvme_admin": false, 00:13:04.402 "nvme_io": false, 00:13:04.402 "nvme_io_md": false, 00:13:04.402 "write_zeroes": true, 00:13:04.402 "zcopy": false, 00:13:04.402 "get_zone_info": false, 00:13:04.402 "zone_management": false, 00:13:04.402 "zone_append": false, 00:13:04.402 "compare": false, 00:13:04.402 "compare_and_write": false, 00:13:04.402 "abort": false, 00:13:04.402 "seek_hole": false, 00:13:04.402 "seek_data": false, 00:13:04.402 "copy": false, 00:13:04.402 "nvme_iov_md": false 00:13:04.402 }, 00:13:04.402 "driver_specific": { 00:13:04.402 "raid": { 00:13:04.402 "uuid": "ab7252b3-69b5-4071-a15f-e1e096f77fef", 00:13:04.402 "strip_size_kb": 64, 00:13:04.402 "state": "online", 00:13:04.402 "raid_level": "raid5f", 00:13:04.402 "superblock": false, 00:13:04.402 "num_base_bdevs": 3, 00:13:04.402 "num_base_bdevs_discovered": 3, 00:13:04.402 "num_base_bdevs_operational": 3, 00:13:04.402 "base_bdevs_list": [ 00:13:04.402 { 00:13:04.402 "name": "BaseBdev1", 00:13:04.402 "uuid": "e071571f-9a6a-4f6b-bc34-ea32ef9f6f7a", 00:13:04.402 "is_configured": true, 00:13:04.402 "data_offset": 0, 00:13:04.402 "data_size": 65536 00:13:04.402 }, 00:13:04.402 { 00:13:04.402 "name": "BaseBdev2", 00:13:04.402 "uuid": "6c900b77-85d4-46ba-85c9-2278c3658f8e", 00:13:04.402 "is_configured": true, 00:13:04.402 "data_offset": 0, 00:13:04.402 "data_size": 65536 00:13:04.402 }, 00:13:04.402 { 00:13:04.402 "name": "BaseBdev3", 00:13:04.402 "uuid": "bb7aae93-c556-464b-b3c1-df18a1e61017", 00:13:04.402 "is_configured": true, 00:13:04.402 "data_offset": 0, 00:13:04.402 "data_size": 65536 00:13:04.402 } 00:13:04.402 ] 00:13:04.402 } 00:13:04.402 } 00:13:04.402 }' 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:04.402 BaseBdev2 00:13:04.402 BaseBdev3' 00:13:04.402 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:04.663 [2024-12-07 05:40:37.944869] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:04.663 "name": "Existed_Raid", 00:13:04.663 "uuid": "ab7252b3-69b5-4071-a15f-e1e096f77fef", 00:13:04.663 "strip_size_kb": 64, 00:13:04.663 "state": "online", 00:13:04.663 "raid_level": "raid5f", 00:13:04.663 "superblock": false, 00:13:04.663 "num_base_bdevs": 3, 00:13:04.663 "num_base_bdevs_discovered": 2, 00:13:04.663 "num_base_bdevs_operational": 2, 00:13:04.663 "base_bdevs_list": [ 00:13:04.663 { 00:13:04.663 "name": null, 00:13:04.663 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:04.663 "is_configured": false, 00:13:04.663 "data_offset": 0, 00:13:04.663 "data_size": 65536 00:13:04.663 }, 00:13:04.663 { 00:13:04.663 "name": "BaseBdev2", 00:13:04.663 "uuid": "6c900b77-85d4-46ba-85c9-2278c3658f8e", 00:13:04.663 "is_configured": true, 00:13:04.663 "data_offset": 0, 00:13:04.663 "data_size": 65536 00:13:04.663 }, 00:13:04.663 { 00:13:04.663 "name": "BaseBdev3", 00:13:04.663 "uuid": "bb7aae93-c556-464b-b3c1-df18a1e61017", 00:13:04.663 "is_configured": true, 00:13:04.663 "data_offset": 0, 00:13:04.663 "data_size": 65536 00:13:04.663 } 00:13:04.663 ] 00:13:04.663 }' 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:04.663 05:40:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.233 [2024-12-07 05:40:38.403249] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:05.233 [2024-12-07 05:40:38.403343] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:05.233 [2024-12-07 05:40:38.414550] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.233 [2024-12-07 05:40:38.474566] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:05.233 [2024-12-07 05:40:38.474608] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:05.233 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.234 BaseBdev2 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.234 [ 00:13:05.234 { 00:13:05.234 "name": "BaseBdev2", 00:13:05.234 "aliases": [ 00:13:05.234 "a3c68229-d177-4930-9a51-ca363a4664cc" 00:13:05.234 ], 00:13:05.234 "product_name": "Malloc disk", 00:13:05.234 "block_size": 512, 00:13:05.234 "num_blocks": 65536, 00:13:05.234 "uuid": "a3c68229-d177-4930-9a51-ca363a4664cc", 00:13:05.234 "assigned_rate_limits": { 00:13:05.234 "rw_ios_per_sec": 0, 00:13:05.234 "rw_mbytes_per_sec": 0, 00:13:05.234 "r_mbytes_per_sec": 0, 00:13:05.234 "w_mbytes_per_sec": 0 00:13:05.234 }, 00:13:05.234 "claimed": false, 00:13:05.234 "zoned": false, 00:13:05.234 "supported_io_types": { 00:13:05.234 "read": true, 00:13:05.234 "write": true, 00:13:05.234 "unmap": true, 00:13:05.234 "flush": true, 00:13:05.234 "reset": true, 00:13:05.234 "nvme_admin": false, 00:13:05.234 "nvme_io": false, 00:13:05.234 "nvme_io_md": false, 00:13:05.234 "write_zeroes": true, 00:13:05.234 "zcopy": true, 00:13:05.234 "get_zone_info": false, 00:13:05.234 "zone_management": false, 00:13:05.234 "zone_append": false, 00:13:05.234 "compare": false, 00:13:05.234 "compare_and_write": false, 00:13:05.234 "abort": true, 00:13:05.234 "seek_hole": false, 00:13:05.234 "seek_data": false, 00:13:05.234 "copy": true, 00:13:05.234 "nvme_iov_md": false 00:13:05.234 }, 00:13:05.234 "memory_domains": [ 00:13:05.234 { 00:13:05.234 "dma_device_id": "system", 00:13:05.234 "dma_device_type": 1 00:13:05.234 }, 00:13:05.234 { 00:13:05.234 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:05.234 "dma_device_type": 2 00:13:05.234 } 00:13:05.234 ], 00:13:05.234 "driver_specific": {} 00:13:05.234 } 00:13:05.234 ] 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.234 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.494 BaseBdev3 00:13:05.494 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.494 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:05.494 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:05.494 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:05.494 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:05.494 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:05.494 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:05.494 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:05.494 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.495 [ 00:13:05.495 { 00:13:05.495 "name": "BaseBdev3", 00:13:05.495 "aliases": [ 00:13:05.495 "e2ea0322-4e29-42aa-afe3-949075479dd3" 00:13:05.495 ], 00:13:05.495 "product_name": "Malloc disk", 00:13:05.495 "block_size": 512, 00:13:05.495 "num_blocks": 65536, 00:13:05.495 "uuid": "e2ea0322-4e29-42aa-afe3-949075479dd3", 00:13:05.495 "assigned_rate_limits": { 00:13:05.495 "rw_ios_per_sec": 0, 00:13:05.495 "rw_mbytes_per_sec": 0, 00:13:05.495 "r_mbytes_per_sec": 0, 00:13:05.495 "w_mbytes_per_sec": 0 00:13:05.495 }, 00:13:05.495 "claimed": false, 00:13:05.495 "zoned": false, 00:13:05.495 "supported_io_types": { 00:13:05.495 "read": true, 00:13:05.495 "write": true, 00:13:05.495 "unmap": true, 00:13:05.495 "flush": true, 00:13:05.495 "reset": true, 00:13:05.495 "nvme_admin": false, 00:13:05.495 "nvme_io": false, 00:13:05.495 "nvme_io_md": false, 00:13:05.495 "write_zeroes": true, 00:13:05.495 "zcopy": true, 00:13:05.495 "get_zone_info": false, 00:13:05.495 "zone_management": false, 00:13:05.495 "zone_append": false, 00:13:05.495 "compare": false, 00:13:05.495 "compare_and_write": false, 00:13:05.495 "abort": true, 00:13:05.495 "seek_hole": false, 00:13:05.495 "seek_data": false, 00:13:05.495 "copy": true, 00:13:05.495 "nvme_iov_md": false 00:13:05.495 }, 00:13:05.495 "memory_domains": [ 00:13:05.495 { 00:13:05.495 "dma_device_id": "system", 00:13:05.495 "dma_device_type": 1 00:13:05.495 }, 00:13:05.495 { 00:13:05.495 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:05.495 "dma_device_type": 2 00:13:05.495 } 00:13:05.495 ], 00:13:05.495 "driver_specific": {} 00:13:05.495 } 00:13:05.495 ] 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.495 [2024-12-07 05:40:38.649511] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:05.495 [2024-12-07 05:40:38.649603] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:05.495 [2024-12-07 05:40:38.649662] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:05.495 [2024-12-07 05:40:38.651454] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:05.495 "name": "Existed_Raid", 00:13:05.495 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.495 "strip_size_kb": 64, 00:13:05.495 "state": "configuring", 00:13:05.495 "raid_level": "raid5f", 00:13:05.495 "superblock": false, 00:13:05.495 "num_base_bdevs": 3, 00:13:05.495 "num_base_bdevs_discovered": 2, 00:13:05.495 "num_base_bdevs_operational": 3, 00:13:05.495 "base_bdevs_list": [ 00:13:05.495 { 00:13:05.495 "name": "BaseBdev1", 00:13:05.495 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.495 "is_configured": false, 00:13:05.495 "data_offset": 0, 00:13:05.495 "data_size": 0 00:13:05.495 }, 00:13:05.495 { 00:13:05.495 "name": "BaseBdev2", 00:13:05.495 "uuid": "a3c68229-d177-4930-9a51-ca363a4664cc", 00:13:05.495 "is_configured": true, 00:13:05.495 "data_offset": 0, 00:13:05.495 "data_size": 65536 00:13:05.495 }, 00:13:05.495 { 00:13:05.495 "name": "BaseBdev3", 00:13:05.495 "uuid": "e2ea0322-4e29-42aa-afe3-949075479dd3", 00:13:05.495 "is_configured": true, 00:13:05.495 "data_offset": 0, 00:13:05.495 "data_size": 65536 00:13:05.495 } 00:13:05.495 ] 00:13:05.495 }' 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:05.495 05:40:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.755 [2024-12-07 05:40:39.072765] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:05.755 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.015 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:06.015 "name": "Existed_Raid", 00:13:06.015 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.015 "strip_size_kb": 64, 00:13:06.015 "state": "configuring", 00:13:06.015 "raid_level": "raid5f", 00:13:06.015 "superblock": false, 00:13:06.015 "num_base_bdevs": 3, 00:13:06.015 "num_base_bdevs_discovered": 1, 00:13:06.015 "num_base_bdevs_operational": 3, 00:13:06.015 "base_bdevs_list": [ 00:13:06.015 { 00:13:06.015 "name": "BaseBdev1", 00:13:06.015 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.015 "is_configured": false, 00:13:06.015 "data_offset": 0, 00:13:06.015 "data_size": 0 00:13:06.015 }, 00:13:06.015 { 00:13:06.015 "name": null, 00:13:06.015 "uuid": "a3c68229-d177-4930-9a51-ca363a4664cc", 00:13:06.015 "is_configured": false, 00:13:06.015 "data_offset": 0, 00:13:06.015 "data_size": 65536 00:13:06.015 }, 00:13:06.015 { 00:13:06.015 "name": "BaseBdev3", 00:13:06.015 "uuid": "e2ea0322-4e29-42aa-afe3-949075479dd3", 00:13:06.015 "is_configured": true, 00:13:06.015 "data_offset": 0, 00:13:06.015 "data_size": 65536 00:13:06.015 } 00:13:06.015 ] 00:13:06.015 }' 00:13:06.015 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:06.015 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.275 [2024-12-07 05:40:39.562922] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:06.275 BaseBdev1 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.275 [ 00:13:06.275 { 00:13:06.275 "name": "BaseBdev1", 00:13:06.275 "aliases": [ 00:13:06.275 "5653f420-fa67-4a2e-a4ca-a12473690f29" 00:13:06.275 ], 00:13:06.275 "product_name": "Malloc disk", 00:13:06.275 "block_size": 512, 00:13:06.275 "num_blocks": 65536, 00:13:06.275 "uuid": "5653f420-fa67-4a2e-a4ca-a12473690f29", 00:13:06.275 "assigned_rate_limits": { 00:13:06.275 "rw_ios_per_sec": 0, 00:13:06.275 "rw_mbytes_per_sec": 0, 00:13:06.275 "r_mbytes_per_sec": 0, 00:13:06.275 "w_mbytes_per_sec": 0 00:13:06.275 }, 00:13:06.275 "claimed": true, 00:13:06.275 "claim_type": "exclusive_write", 00:13:06.275 "zoned": false, 00:13:06.275 "supported_io_types": { 00:13:06.275 "read": true, 00:13:06.275 "write": true, 00:13:06.275 "unmap": true, 00:13:06.275 "flush": true, 00:13:06.275 "reset": true, 00:13:06.275 "nvme_admin": false, 00:13:06.275 "nvme_io": false, 00:13:06.275 "nvme_io_md": false, 00:13:06.275 "write_zeroes": true, 00:13:06.275 "zcopy": true, 00:13:06.275 "get_zone_info": false, 00:13:06.275 "zone_management": false, 00:13:06.275 "zone_append": false, 00:13:06.275 "compare": false, 00:13:06.275 "compare_and_write": false, 00:13:06.275 "abort": true, 00:13:06.275 "seek_hole": false, 00:13:06.275 "seek_data": false, 00:13:06.275 "copy": true, 00:13:06.275 "nvme_iov_md": false 00:13:06.275 }, 00:13:06.275 "memory_domains": [ 00:13:06.275 { 00:13:06.275 "dma_device_id": "system", 00:13:06.275 "dma_device_type": 1 00:13:06.275 }, 00:13:06.275 { 00:13:06.275 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:06.275 "dma_device_type": 2 00:13:06.275 } 00:13:06.275 ], 00:13:06.275 "driver_specific": {} 00:13:06.275 } 00:13:06.275 ] 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.275 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.535 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:06.535 "name": "Existed_Raid", 00:13:06.535 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.535 "strip_size_kb": 64, 00:13:06.535 "state": "configuring", 00:13:06.535 "raid_level": "raid5f", 00:13:06.535 "superblock": false, 00:13:06.535 "num_base_bdevs": 3, 00:13:06.535 "num_base_bdevs_discovered": 2, 00:13:06.535 "num_base_bdevs_operational": 3, 00:13:06.535 "base_bdevs_list": [ 00:13:06.535 { 00:13:06.535 "name": "BaseBdev1", 00:13:06.535 "uuid": "5653f420-fa67-4a2e-a4ca-a12473690f29", 00:13:06.535 "is_configured": true, 00:13:06.535 "data_offset": 0, 00:13:06.535 "data_size": 65536 00:13:06.535 }, 00:13:06.535 { 00:13:06.535 "name": null, 00:13:06.535 "uuid": "a3c68229-d177-4930-9a51-ca363a4664cc", 00:13:06.535 "is_configured": false, 00:13:06.535 "data_offset": 0, 00:13:06.535 "data_size": 65536 00:13:06.535 }, 00:13:06.535 { 00:13:06.535 "name": "BaseBdev3", 00:13:06.535 "uuid": "e2ea0322-4e29-42aa-afe3-949075479dd3", 00:13:06.535 "is_configured": true, 00:13:06.535 "data_offset": 0, 00:13:06.535 "data_size": 65536 00:13:06.535 } 00:13:06.535 ] 00:13:06.535 }' 00:13:06.535 05:40:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:06.535 05:40:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.793 [2024-12-07 05:40:40.034171] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:06.793 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:06.794 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:06.794 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:06.794 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:06.794 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.794 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.794 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.794 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:06.794 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.794 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:06.794 "name": "Existed_Raid", 00:13:06.794 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.794 "strip_size_kb": 64, 00:13:06.794 "state": "configuring", 00:13:06.794 "raid_level": "raid5f", 00:13:06.794 "superblock": false, 00:13:06.794 "num_base_bdevs": 3, 00:13:06.794 "num_base_bdevs_discovered": 1, 00:13:06.794 "num_base_bdevs_operational": 3, 00:13:06.794 "base_bdevs_list": [ 00:13:06.794 { 00:13:06.794 "name": "BaseBdev1", 00:13:06.794 "uuid": "5653f420-fa67-4a2e-a4ca-a12473690f29", 00:13:06.794 "is_configured": true, 00:13:06.794 "data_offset": 0, 00:13:06.794 "data_size": 65536 00:13:06.794 }, 00:13:06.794 { 00:13:06.794 "name": null, 00:13:06.794 "uuid": "a3c68229-d177-4930-9a51-ca363a4664cc", 00:13:06.794 "is_configured": false, 00:13:06.794 "data_offset": 0, 00:13:06.794 "data_size": 65536 00:13:06.794 }, 00:13:06.794 { 00:13:06.794 "name": null, 00:13:06.794 "uuid": "e2ea0322-4e29-42aa-afe3-949075479dd3", 00:13:06.794 "is_configured": false, 00:13:06.794 "data_offset": 0, 00:13:06.794 "data_size": 65536 00:13:06.794 } 00:13:06.794 ] 00:13:06.794 }' 00:13:06.794 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:06.794 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:07.362 [2024-12-07 05:40:40.509387] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:07.362 "name": "Existed_Raid", 00:13:07.362 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.362 "strip_size_kb": 64, 00:13:07.362 "state": "configuring", 00:13:07.362 "raid_level": "raid5f", 00:13:07.362 "superblock": false, 00:13:07.362 "num_base_bdevs": 3, 00:13:07.362 "num_base_bdevs_discovered": 2, 00:13:07.362 "num_base_bdevs_operational": 3, 00:13:07.362 "base_bdevs_list": [ 00:13:07.362 { 00:13:07.362 "name": "BaseBdev1", 00:13:07.362 "uuid": "5653f420-fa67-4a2e-a4ca-a12473690f29", 00:13:07.362 "is_configured": true, 00:13:07.362 "data_offset": 0, 00:13:07.362 "data_size": 65536 00:13:07.362 }, 00:13:07.362 { 00:13:07.362 "name": null, 00:13:07.362 "uuid": "a3c68229-d177-4930-9a51-ca363a4664cc", 00:13:07.362 "is_configured": false, 00:13:07.362 "data_offset": 0, 00:13:07.362 "data_size": 65536 00:13:07.362 }, 00:13:07.362 { 00:13:07.362 "name": "BaseBdev3", 00:13:07.362 "uuid": "e2ea0322-4e29-42aa-afe3-949075479dd3", 00:13:07.362 "is_configured": true, 00:13:07.362 "data_offset": 0, 00:13:07.362 "data_size": 65536 00:13:07.362 } 00:13:07.362 ] 00:13:07.362 }' 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:07.362 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:07.621 [2024-12-07 05:40:40.972606] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:07.621 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:07.880 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.880 05:40:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:07.880 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.880 05:40:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:07.880 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.880 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:07.880 "name": "Existed_Raid", 00:13:07.880 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.880 "strip_size_kb": 64, 00:13:07.880 "state": "configuring", 00:13:07.880 "raid_level": "raid5f", 00:13:07.880 "superblock": false, 00:13:07.880 "num_base_bdevs": 3, 00:13:07.880 "num_base_bdevs_discovered": 1, 00:13:07.880 "num_base_bdevs_operational": 3, 00:13:07.880 "base_bdevs_list": [ 00:13:07.880 { 00:13:07.880 "name": null, 00:13:07.880 "uuid": "5653f420-fa67-4a2e-a4ca-a12473690f29", 00:13:07.880 "is_configured": false, 00:13:07.880 "data_offset": 0, 00:13:07.880 "data_size": 65536 00:13:07.880 }, 00:13:07.880 { 00:13:07.880 "name": null, 00:13:07.880 "uuid": "a3c68229-d177-4930-9a51-ca363a4664cc", 00:13:07.880 "is_configured": false, 00:13:07.880 "data_offset": 0, 00:13:07.880 "data_size": 65536 00:13:07.880 }, 00:13:07.880 { 00:13:07.880 "name": "BaseBdev3", 00:13:07.880 "uuid": "e2ea0322-4e29-42aa-afe3-949075479dd3", 00:13:07.880 "is_configured": true, 00:13:07.880 "data_offset": 0, 00:13:07.880 "data_size": 65536 00:13:07.880 } 00:13:07.880 ] 00:13:07.880 }' 00:13:07.880 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:07.880 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.139 [2024-12-07 05:40:41.450144] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:08.139 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.399 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:08.399 "name": "Existed_Raid", 00:13:08.399 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.399 "strip_size_kb": 64, 00:13:08.399 "state": "configuring", 00:13:08.399 "raid_level": "raid5f", 00:13:08.399 "superblock": false, 00:13:08.399 "num_base_bdevs": 3, 00:13:08.399 "num_base_bdevs_discovered": 2, 00:13:08.399 "num_base_bdevs_operational": 3, 00:13:08.399 "base_bdevs_list": [ 00:13:08.399 { 00:13:08.399 "name": null, 00:13:08.399 "uuid": "5653f420-fa67-4a2e-a4ca-a12473690f29", 00:13:08.399 "is_configured": false, 00:13:08.399 "data_offset": 0, 00:13:08.399 "data_size": 65536 00:13:08.399 }, 00:13:08.399 { 00:13:08.399 "name": "BaseBdev2", 00:13:08.399 "uuid": "a3c68229-d177-4930-9a51-ca363a4664cc", 00:13:08.399 "is_configured": true, 00:13:08.399 "data_offset": 0, 00:13:08.399 "data_size": 65536 00:13:08.399 }, 00:13:08.399 { 00:13:08.399 "name": "BaseBdev3", 00:13:08.399 "uuid": "e2ea0322-4e29-42aa-afe3-949075479dd3", 00:13:08.399 "is_configured": true, 00:13:08.399 "data_offset": 0, 00:13:08.399 "data_size": 65536 00:13:08.399 } 00:13:08.399 ] 00:13:08.399 }' 00:13:08.399 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:08.399 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 5653f420-fa67-4a2e-a4ca-a12473690f29 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.658 [2024-12-07 05:40:41.992106] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:08.658 [2024-12-07 05:40:41.992149] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:08.658 [2024-12-07 05:40:41.992159] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:08.658 [2024-12-07 05:40:41.992396] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:08.658 [2024-12-07 05:40:41.992811] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:08.658 [2024-12-07 05:40:41.992823] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:13:08.658 [2024-12-07 05:40:41.993006] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:08.658 NewBaseBdev 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.658 05:40:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.658 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.658 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:08.658 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.658 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.658 [ 00:13:08.658 { 00:13:08.658 "name": "NewBaseBdev", 00:13:08.658 "aliases": [ 00:13:08.658 "5653f420-fa67-4a2e-a4ca-a12473690f29" 00:13:08.658 ], 00:13:08.658 "product_name": "Malloc disk", 00:13:08.658 "block_size": 512, 00:13:08.658 "num_blocks": 65536, 00:13:08.658 "uuid": "5653f420-fa67-4a2e-a4ca-a12473690f29", 00:13:08.658 "assigned_rate_limits": { 00:13:08.658 "rw_ios_per_sec": 0, 00:13:08.658 "rw_mbytes_per_sec": 0, 00:13:08.658 "r_mbytes_per_sec": 0, 00:13:08.658 "w_mbytes_per_sec": 0 00:13:08.658 }, 00:13:08.658 "claimed": true, 00:13:08.658 "claim_type": "exclusive_write", 00:13:08.658 "zoned": false, 00:13:08.658 "supported_io_types": { 00:13:08.658 "read": true, 00:13:08.658 "write": true, 00:13:08.658 "unmap": true, 00:13:08.658 "flush": true, 00:13:08.658 "reset": true, 00:13:08.658 "nvme_admin": false, 00:13:08.917 "nvme_io": false, 00:13:08.917 "nvme_io_md": false, 00:13:08.917 "write_zeroes": true, 00:13:08.917 "zcopy": true, 00:13:08.917 "get_zone_info": false, 00:13:08.917 "zone_management": false, 00:13:08.917 "zone_append": false, 00:13:08.917 "compare": false, 00:13:08.917 "compare_and_write": false, 00:13:08.917 "abort": true, 00:13:08.917 "seek_hole": false, 00:13:08.917 "seek_data": false, 00:13:08.917 "copy": true, 00:13:08.917 "nvme_iov_md": false 00:13:08.917 }, 00:13:08.917 "memory_domains": [ 00:13:08.917 { 00:13:08.917 "dma_device_id": "system", 00:13:08.917 "dma_device_type": 1 00:13:08.917 }, 00:13:08.917 { 00:13:08.917 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:08.917 "dma_device_type": 2 00:13:08.917 } 00:13:08.917 ], 00:13:08.917 "driver_specific": {} 00:13:08.917 } 00:13:08.917 ] 00:13:08.917 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.917 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:08.917 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:08.918 "name": "Existed_Raid", 00:13:08.918 "uuid": "3cc3fa7c-d863-4b6e-9ee9-9b9de7db1606", 00:13:08.918 "strip_size_kb": 64, 00:13:08.918 "state": "online", 00:13:08.918 "raid_level": "raid5f", 00:13:08.918 "superblock": false, 00:13:08.918 "num_base_bdevs": 3, 00:13:08.918 "num_base_bdevs_discovered": 3, 00:13:08.918 "num_base_bdevs_operational": 3, 00:13:08.918 "base_bdevs_list": [ 00:13:08.918 { 00:13:08.918 "name": "NewBaseBdev", 00:13:08.918 "uuid": "5653f420-fa67-4a2e-a4ca-a12473690f29", 00:13:08.918 "is_configured": true, 00:13:08.918 "data_offset": 0, 00:13:08.918 "data_size": 65536 00:13:08.918 }, 00:13:08.918 { 00:13:08.918 "name": "BaseBdev2", 00:13:08.918 "uuid": "a3c68229-d177-4930-9a51-ca363a4664cc", 00:13:08.918 "is_configured": true, 00:13:08.918 "data_offset": 0, 00:13:08.918 "data_size": 65536 00:13:08.918 }, 00:13:08.918 { 00:13:08.918 "name": "BaseBdev3", 00:13:08.918 "uuid": "e2ea0322-4e29-42aa-afe3-949075479dd3", 00:13:08.918 "is_configured": true, 00:13:08.918 "data_offset": 0, 00:13:08.918 "data_size": 65536 00:13:08.918 } 00:13:08.918 ] 00:13:08.918 }' 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:08.918 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.177 [2024-12-07 05:40:42.439557] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:09.177 "name": "Existed_Raid", 00:13:09.177 "aliases": [ 00:13:09.177 "3cc3fa7c-d863-4b6e-9ee9-9b9de7db1606" 00:13:09.177 ], 00:13:09.177 "product_name": "Raid Volume", 00:13:09.177 "block_size": 512, 00:13:09.177 "num_blocks": 131072, 00:13:09.177 "uuid": "3cc3fa7c-d863-4b6e-9ee9-9b9de7db1606", 00:13:09.177 "assigned_rate_limits": { 00:13:09.177 "rw_ios_per_sec": 0, 00:13:09.177 "rw_mbytes_per_sec": 0, 00:13:09.177 "r_mbytes_per_sec": 0, 00:13:09.177 "w_mbytes_per_sec": 0 00:13:09.177 }, 00:13:09.177 "claimed": false, 00:13:09.177 "zoned": false, 00:13:09.177 "supported_io_types": { 00:13:09.177 "read": true, 00:13:09.177 "write": true, 00:13:09.177 "unmap": false, 00:13:09.177 "flush": false, 00:13:09.177 "reset": true, 00:13:09.177 "nvme_admin": false, 00:13:09.177 "nvme_io": false, 00:13:09.177 "nvme_io_md": false, 00:13:09.177 "write_zeroes": true, 00:13:09.177 "zcopy": false, 00:13:09.177 "get_zone_info": false, 00:13:09.177 "zone_management": false, 00:13:09.177 "zone_append": false, 00:13:09.177 "compare": false, 00:13:09.177 "compare_and_write": false, 00:13:09.177 "abort": false, 00:13:09.177 "seek_hole": false, 00:13:09.177 "seek_data": false, 00:13:09.177 "copy": false, 00:13:09.177 "nvme_iov_md": false 00:13:09.177 }, 00:13:09.177 "driver_specific": { 00:13:09.177 "raid": { 00:13:09.177 "uuid": "3cc3fa7c-d863-4b6e-9ee9-9b9de7db1606", 00:13:09.177 "strip_size_kb": 64, 00:13:09.177 "state": "online", 00:13:09.177 "raid_level": "raid5f", 00:13:09.177 "superblock": false, 00:13:09.177 "num_base_bdevs": 3, 00:13:09.177 "num_base_bdevs_discovered": 3, 00:13:09.177 "num_base_bdevs_operational": 3, 00:13:09.177 "base_bdevs_list": [ 00:13:09.177 { 00:13:09.177 "name": "NewBaseBdev", 00:13:09.177 "uuid": "5653f420-fa67-4a2e-a4ca-a12473690f29", 00:13:09.177 "is_configured": true, 00:13:09.177 "data_offset": 0, 00:13:09.177 "data_size": 65536 00:13:09.177 }, 00:13:09.177 { 00:13:09.177 "name": "BaseBdev2", 00:13:09.177 "uuid": "a3c68229-d177-4930-9a51-ca363a4664cc", 00:13:09.177 "is_configured": true, 00:13:09.177 "data_offset": 0, 00:13:09.177 "data_size": 65536 00:13:09.177 }, 00:13:09.177 { 00:13:09.177 "name": "BaseBdev3", 00:13:09.177 "uuid": "e2ea0322-4e29-42aa-afe3-949075479dd3", 00:13:09.177 "is_configured": true, 00:13:09.177 "data_offset": 0, 00:13:09.177 "data_size": 65536 00:13:09.177 } 00:13:09.177 ] 00:13:09.177 } 00:13:09.177 } 00:13:09.177 }' 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:09.177 BaseBdev2 00:13:09.177 BaseBdev3' 00:13:09.177 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.436 [2024-12-07 05:40:42.698918] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:09.436 [2024-12-07 05:40:42.698942] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:09.436 [2024-12-07 05:40:42.699003] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:09.436 [2024-12-07 05:40:42.699239] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:09.436 [2024-12-07 05:40:42.699252] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 90071 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 90071 ']' 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 90071 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 90071 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 90071' 00:13:09.436 killing process with pid 90071 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 90071 00:13:09.436 [2024-12-07 05:40:42.744193] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:09.436 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 90071 00:13:09.436 [2024-12-07 05:40:42.775566] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:09.695 05:40:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:13:09.695 00:13:09.695 real 0m8.491s 00:13:09.695 user 0m14.526s 00:13:09.695 sys 0m1.710s 00:13:09.695 ************************************ 00:13:09.695 END TEST raid5f_state_function_test 00:13:09.695 ************************************ 00:13:09.695 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:09.695 05:40:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.695 05:40:43 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:13:09.695 05:40:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:13:09.695 05:40:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:09.695 05:40:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:09.954 ************************************ 00:13:09.954 START TEST raid5f_state_function_test_sb 00:13:09.954 ************************************ 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 true 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=90670 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90670' 00:13:09.954 Process raid pid: 90670 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 90670 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 90670 ']' 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:09.954 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:09.954 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:09.954 [2024-12-07 05:40:43.155559] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:13:09.954 [2024-12-07 05:40:43.155774] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:09.954 [2024-12-07 05:40:43.310900] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:10.212 [2024-12-07 05:40:43.335446] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:10.212 [2024-12-07 05:40:43.377361] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:10.212 [2024-12-07 05:40:43.377485] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:10.778 [2024-12-07 05:40:43.983870] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:10.778 [2024-12-07 05:40:43.983997] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:10.778 [2024-12-07 05:40:43.984029] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:10.778 [2024-12-07 05:40:43.984054] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:10.778 [2024-12-07 05:40:43.984072] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:10.778 [2024-12-07 05:40:43.984094] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.778 05:40:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:10.778 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.778 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:10.778 "name": "Existed_Raid", 00:13:10.778 "uuid": "bef0a968-12ee-484b-84e4-276b3eb9a470", 00:13:10.778 "strip_size_kb": 64, 00:13:10.778 "state": "configuring", 00:13:10.778 "raid_level": "raid5f", 00:13:10.778 "superblock": true, 00:13:10.778 "num_base_bdevs": 3, 00:13:10.778 "num_base_bdevs_discovered": 0, 00:13:10.778 "num_base_bdevs_operational": 3, 00:13:10.778 "base_bdevs_list": [ 00:13:10.778 { 00:13:10.778 "name": "BaseBdev1", 00:13:10.778 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.778 "is_configured": false, 00:13:10.778 "data_offset": 0, 00:13:10.778 "data_size": 0 00:13:10.778 }, 00:13:10.778 { 00:13:10.778 "name": "BaseBdev2", 00:13:10.778 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.778 "is_configured": false, 00:13:10.778 "data_offset": 0, 00:13:10.778 "data_size": 0 00:13:10.778 }, 00:13:10.778 { 00:13:10.778 "name": "BaseBdev3", 00:13:10.778 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.778 "is_configured": false, 00:13:10.778 "data_offset": 0, 00:13:10.778 "data_size": 0 00:13:10.778 } 00:13:10.778 ] 00:13:10.778 }' 00:13:10.778 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:10.778 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.038 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:11.038 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.038 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.038 [2024-12-07 05:40:44.399093] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:11.038 [2024-12-07 05:40:44.399179] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:13:11.298 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.298 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:11.298 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.298 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.298 [2024-12-07 05:40:44.411107] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:11.298 [2024-12-07 05:40:44.411151] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:11.298 [2024-12-07 05:40:44.411160] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:11.298 [2024-12-07 05:40:44.411168] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:11.298 [2024-12-07 05:40:44.411174] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:11.299 [2024-12-07 05:40:44.411182] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.299 [2024-12-07 05:40:44.431862] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:11.299 BaseBdev1 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.299 [ 00:13:11.299 { 00:13:11.299 "name": "BaseBdev1", 00:13:11.299 "aliases": [ 00:13:11.299 "9dbe4582-9102-4dc3-abec-5e51365c79b4" 00:13:11.299 ], 00:13:11.299 "product_name": "Malloc disk", 00:13:11.299 "block_size": 512, 00:13:11.299 "num_blocks": 65536, 00:13:11.299 "uuid": "9dbe4582-9102-4dc3-abec-5e51365c79b4", 00:13:11.299 "assigned_rate_limits": { 00:13:11.299 "rw_ios_per_sec": 0, 00:13:11.299 "rw_mbytes_per_sec": 0, 00:13:11.299 "r_mbytes_per_sec": 0, 00:13:11.299 "w_mbytes_per_sec": 0 00:13:11.299 }, 00:13:11.299 "claimed": true, 00:13:11.299 "claim_type": "exclusive_write", 00:13:11.299 "zoned": false, 00:13:11.299 "supported_io_types": { 00:13:11.299 "read": true, 00:13:11.299 "write": true, 00:13:11.299 "unmap": true, 00:13:11.299 "flush": true, 00:13:11.299 "reset": true, 00:13:11.299 "nvme_admin": false, 00:13:11.299 "nvme_io": false, 00:13:11.299 "nvme_io_md": false, 00:13:11.299 "write_zeroes": true, 00:13:11.299 "zcopy": true, 00:13:11.299 "get_zone_info": false, 00:13:11.299 "zone_management": false, 00:13:11.299 "zone_append": false, 00:13:11.299 "compare": false, 00:13:11.299 "compare_and_write": false, 00:13:11.299 "abort": true, 00:13:11.299 "seek_hole": false, 00:13:11.299 "seek_data": false, 00:13:11.299 "copy": true, 00:13:11.299 "nvme_iov_md": false 00:13:11.299 }, 00:13:11.299 "memory_domains": [ 00:13:11.299 { 00:13:11.299 "dma_device_id": "system", 00:13:11.299 "dma_device_type": 1 00:13:11.299 }, 00:13:11.299 { 00:13:11.299 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:11.299 "dma_device_type": 2 00:13:11.299 } 00:13:11.299 ], 00:13:11.299 "driver_specific": {} 00:13:11.299 } 00:13:11.299 ] 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.299 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:11.299 "name": "Existed_Raid", 00:13:11.299 "uuid": "4b6f0d5d-d576-4bcd-a471-b7b0c90f260f", 00:13:11.299 "strip_size_kb": 64, 00:13:11.299 "state": "configuring", 00:13:11.299 "raid_level": "raid5f", 00:13:11.299 "superblock": true, 00:13:11.299 "num_base_bdevs": 3, 00:13:11.299 "num_base_bdevs_discovered": 1, 00:13:11.299 "num_base_bdevs_operational": 3, 00:13:11.299 "base_bdevs_list": [ 00:13:11.299 { 00:13:11.299 "name": "BaseBdev1", 00:13:11.299 "uuid": "9dbe4582-9102-4dc3-abec-5e51365c79b4", 00:13:11.299 "is_configured": true, 00:13:11.300 "data_offset": 2048, 00:13:11.300 "data_size": 63488 00:13:11.300 }, 00:13:11.300 { 00:13:11.300 "name": "BaseBdev2", 00:13:11.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.300 "is_configured": false, 00:13:11.300 "data_offset": 0, 00:13:11.300 "data_size": 0 00:13:11.300 }, 00:13:11.300 { 00:13:11.300 "name": "BaseBdev3", 00:13:11.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.300 "is_configured": false, 00:13:11.300 "data_offset": 0, 00:13:11.300 "data_size": 0 00:13:11.300 } 00:13:11.300 ] 00:13:11.300 }' 00:13:11.300 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:11.300 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.872 [2024-12-07 05:40:44.939062] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:11.872 [2024-12-07 05:40:44.939158] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.872 [2024-12-07 05:40:44.947090] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:11.872 [2024-12-07 05:40:44.948918] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:11.872 [2024-12-07 05:40:44.948996] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:11.872 [2024-12-07 05:40:44.949024] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:11.872 [2024-12-07 05:40:44.949048] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:11.872 "name": "Existed_Raid", 00:13:11.872 "uuid": "8ce50fc8-8335-42ba-833f-27b47b083a20", 00:13:11.872 "strip_size_kb": 64, 00:13:11.872 "state": "configuring", 00:13:11.872 "raid_level": "raid5f", 00:13:11.872 "superblock": true, 00:13:11.872 "num_base_bdevs": 3, 00:13:11.872 "num_base_bdevs_discovered": 1, 00:13:11.872 "num_base_bdevs_operational": 3, 00:13:11.872 "base_bdevs_list": [ 00:13:11.872 { 00:13:11.872 "name": "BaseBdev1", 00:13:11.872 "uuid": "9dbe4582-9102-4dc3-abec-5e51365c79b4", 00:13:11.872 "is_configured": true, 00:13:11.872 "data_offset": 2048, 00:13:11.872 "data_size": 63488 00:13:11.872 }, 00:13:11.872 { 00:13:11.872 "name": "BaseBdev2", 00:13:11.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.872 "is_configured": false, 00:13:11.872 "data_offset": 0, 00:13:11.872 "data_size": 0 00:13:11.872 }, 00:13:11.872 { 00:13:11.872 "name": "BaseBdev3", 00:13:11.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.872 "is_configured": false, 00:13:11.872 "data_offset": 0, 00:13:11.872 "data_size": 0 00:13:11.872 } 00:13:11.872 ] 00:13:11.872 }' 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:11.872 05:40:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.140 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.141 [2024-12-07 05:40:45.413144] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:12.141 BaseBdev2 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.141 [ 00:13:12.141 { 00:13:12.141 "name": "BaseBdev2", 00:13:12.141 "aliases": [ 00:13:12.141 "81f8add4-c768-4ffb-9430-da8cf6fcf2a5" 00:13:12.141 ], 00:13:12.141 "product_name": "Malloc disk", 00:13:12.141 "block_size": 512, 00:13:12.141 "num_blocks": 65536, 00:13:12.141 "uuid": "81f8add4-c768-4ffb-9430-da8cf6fcf2a5", 00:13:12.141 "assigned_rate_limits": { 00:13:12.141 "rw_ios_per_sec": 0, 00:13:12.141 "rw_mbytes_per_sec": 0, 00:13:12.141 "r_mbytes_per_sec": 0, 00:13:12.141 "w_mbytes_per_sec": 0 00:13:12.141 }, 00:13:12.141 "claimed": true, 00:13:12.141 "claim_type": "exclusive_write", 00:13:12.141 "zoned": false, 00:13:12.141 "supported_io_types": { 00:13:12.141 "read": true, 00:13:12.141 "write": true, 00:13:12.141 "unmap": true, 00:13:12.141 "flush": true, 00:13:12.141 "reset": true, 00:13:12.141 "nvme_admin": false, 00:13:12.141 "nvme_io": false, 00:13:12.141 "nvme_io_md": false, 00:13:12.141 "write_zeroes": true, 00:13:12.141 "zcopy": true, 00:13:12.141 "get_zone_info": false, 00:13:12.141 "zone_management": false, 00:13:12.141 "zone_append": false, 00:13:12.141 "compare": false, 00:13:12.141 "compare_and_write": false, 00:13:12.141 "abort": true, 00:13:12.141 "seek_hole": false, 00:13:12.141 "seek_data": false, 00:13:12.141 "copy": true, 00:13:12.141 "nvme_iov_md": false 00:13:12.141 }, 00:13:12.141 "memory_domains": [ 00:13:12.141 { 00:13:12.141 "dma_device_id": "system", 00:13:12.141 "dma_device_type": 1 00:13:12.141 }, 00:13:12.141 { 00:13:12.141 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:12.141 "dma_device_type": 2 00:13:12.141 } 00:13:12.141 ], 00:13:12.141 "driver_specific": {} 00:13:12.141 } 00:13:12.141 ] 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.141 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:12.141 "name": "Existed_Raid", 00:13:12.141 "uuid": "8ce50fc8-8335-42ba-833f-27b47b083a20", 00:13:12.141 "strip_size_kb": 64, 00:13:12.141 "state": "configuring", 00:13:12.141 "raid_level": "raid5f", 00:13:12.141 "superblock": true, 00:13:12.141 "num_base_bdevs": 3, 00:13:12.141 "num_base_bdevs_discovered": 2, 00:13:12.141 "num_base_bdevs_operational": 3, 00:13:12.142 "base_bdevs_list": [ 00:13:12.142 { 00:13:12.142 "name": "BaseBdev1", 00:13:12.142 "uuid": "9dbe4582-9102-4dc3-abec-5e51365c79b4", 00:13:12.142 "is_configured": true, 00:13:12.142 "data_offset": 2048, 00:13:12.142 "data_size": 63488 00:13:12.142 }, 00:13:12.142 { 00:13:12.142 "name": "BaseBdev2", 00:13:12.142 "uuid": "81f8add4-c768-4ffb-9430-da8cf6fcf2a5", 00:13:12.142 "is_configured": true, 00:13:12.142 "data_offset": 2048, 00:13:12.142 "data_size": 63488 00:13:12.142 }, 00:13:12.142 { 00:13:12.142 "name": "BaseBdev3", 00:13:12.142 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:12.142 "is_configured": false, 00:13:12.142 "data_offset": 0, 00:13:12.142 "data_size": 0 00:13:12.142 } 00:13:12.142 ] 00:13:12.142 }' 00:13:12.142 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:12.142 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.711 [2024-12-07 05:40:45.882727] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:12.711 [2024-12-07 05:40:45.883041] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:12.711 [2024-12-07 05:40:45.883069] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:12.711 BaseBdev3 00:13:12.711 [2024-12-07 05:40:45.883485] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:13:12.711 [2024-12-07 05:40:45.884208] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:12.711 [2024-12-07 05:40:45.884235] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.711 [2024-12-07 05:40:45.884419] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.711 [ 00:13:12.711 { 00:13:12.711 "name": "BaseBdev3", 00:13:12.711 "aliases": [ 00:13:12.711 "fb1d39bb-0e6e-485f-aabf-9ece41af1f40" 00:13:12.711 ], 00:13:12.711 "product_name": "Malloc disk", 00:13:12.711 "block_size": 512, 00:13:12.711 "num_blocks": 65536, 00:13:12.711 "uuid": "fb1d39bb-0e6e-485f-aabf-9ece41af1f40", 00:13:12.711 "assigned_rate_limits": { 00:13:12.711 "rw_ios_per_sec": 0, 00:13:12.711 "rw_mbytes_per_sec": 0, 00:13:12.711 "r_mbytes_per_sec": 0, 00:13:12.711 "w_mbytes_per_sec": 0 00:13:12.711 }, 00:13:12.711 "claimed": true, 00:13:12.711 "claim_type": "exclusive_write", 00:13:12.711 "zoned": false, 00:13:12.711 "supported_io_types": { 00:13:12.711 "read": true, 00:13:12.711 "write": true, 00:13:12.711 "unmap": true, 00:13:12.711 "flush": true, 00:13:12.711 "reset": true, 00:13:12.711 "nvme_admin": false, 00:13:12.711 "nvme_io": false, 00:13:12.711 "nvme_io_md": false, 00:13:12.711 "write_zeroes": true, 00:13:12.711 "zcopy": true, 00:13:12.711 "get_zone_info": false, 00:13:12.711 "zone_management": false, 00:13:12.711 "zone_append": false, 00:13:12.711 "compare": false, 00:13:12.711 "compare_and_write": false, 00:13:12.711 "abort": true, 00:13:12.711 "seek_hole": false, 00:13:12.711 "seek_data": false, 00:13:12.711 "copy": true, 00:13:12.711 "nvme_iov_md": false 00:13:12.711 }, 00:13:12.711 "memory_domains": [ 00:13:12.711 { 00:13:12.711 "dma_device_id": "system", 00:13:12.711 "dma_device_type": 1 00:13:12.711 }, 00:13:12.711 { 00:13:12.711 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:12.711 "dma_device_type": 2 00:13:12.711 } 00:13:12.711 ], 00:13:12.711 "driver_specific": {} 00:13:12.711 } 00:13:12.711 ] 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:12.711 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:12.712 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:12.712 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:12.712 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:12.712 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:12.712 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.712 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.712 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.712 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.712 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:12.712 "name": "Existed_Raid", 00:13:12.712 "uuid": "8ce50fc8-8335-42ba-833f-27b47b083a20", 00:13:12.712 "strip_size_kb": 64, 00:13:12.712 "state": "online", 00:13:12.712 "raid_level": "raid5f", 00:13:12.712 "superblock": true, 00:13:12.712 "num_base_bdevs": 3, 00:13:12.712 "num_base_bdevs_discovered": 3, 00:13:12.712 "num_base_bdevs_operational": 3, 00:13:12.712 "base_bdevs_list": [ 00:13:12.712 { 00:13:12.712 "name": "BaseBdev1", 00:13:12.712 "uuid": "9dbe4582-9102-4dc3-abec-5e51365c79b4", 00:13:12.712 "is_configured": true, 00:13:12.712 "data_offset": 2048, 00:13:12.712 "data_size": 63488 00:13:12.712 }, 00:13:12.712 { 00:13:12.712 "name": "BaseBdev2", 00:13:12.712 "uuid": "81f8add4-c768-4ffb-9430-da8cf6fcf2a5", 00:13:12.712 "is_configured": true, 00:13:12.712 "data_offset": 2048, 00:13:12.712 "data_size": 63488 00:13:12.712 }, 00:13:12.712 { 00:13:12.712 "name": "BaseBdev3", 00:13:12.712 "uuid": "fb1d39bb-0e6e-485f-aabf-9ece41af1f40", 00:13:12.712 "is_configured": true, 00:13:12.712 "data_offset": 2048, 00:13:12.712 "data_size": 63488 00:13:12.712 } 00:13:12.712 ] 00:13:12.712 }' 00:13:12.712 05:40:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:12.712 05:40:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.972 [2024-12-07 05:40:46.306245] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:12.972 "name": "Existed_Raid", 00:13:12.972 "aliases": [ 00:13:12.972 "8ce50fc8-8335-42ba-833f-27b47b083a20" 00:13:12.972 ], 00:13:12.972 "product_name": "Raid Volume", 00:13:12.972 "block_size": 512, 00:13:12.972 "num_blocks": 126976, 00:13:12.972 "uuid": "8ce50fc8-8335-42ba-833f-27b47b083a20", 00:13:12.972 "assigned_rate_limits": { 00:13:12.972 "rw_ios_per_sec": 0, 00:13:12.972 "rw_mbytes_per_sec": 0, 00:13:12.972 "r_mbytes_per_sec": 0, 00:13:12.972 "w_mbytes_per_sec": 0 00:13:12.972 }, 00:13:12.972 "claimed": false, 00:13:12.972 "zoned": false, 00:13:12.972 "supported_io_types": { 00:13:12.972 "read": true, 00:13:12.972 "write": true, 00:13:12.972 "unmap": false, 00:13:12.972 "flush": false, 00:13:12.972 "reset": true, 00:13:12.972 "nvme_admin": false, 00:13:12.972 "nvme_io": false, 00:13:12.972 "nvme_io_md": false, 00:13:12.972 "write_zeroes": true, 00:13:12.972 "zcopy": false, 00:13:12.972 "get_zone_info": false, 00:13:12.972 "zone_management": false, 00:13:12.972 "zone_append": false, 00:13:12.972 "compare": false, 00:13:12.972 "compare_and_write": false, 00:13:12.972 "abort": false, 00:13:12.972 "seek_hole": false, 00:13:12.972 "seek_data": false, 00:13:12.972 "copy": false, 00:13:12.972 "nvme_iov_md": false 00:13:12.972 }, 00:13:12.972 "driver_specific": { 00:13:12.972 "raid": { 00:13:12.972 "uuid": "8ce50fc8-8335-42ba-833f-27b47b083a20", 00:13:12.972 "strip_size_kb": 64, 00:13:12.972 "state": "online", 00:13:12.972 "raid_level": "raid5f", 00:13:12.972 "superblock": true, 00:13:12.972 "num_base_bdevs": 3, 00:13:12.972 "num_base_bdevs_discovered": 3, 00:13:12.972 "num_base_bdevs_operational": 3, 00:13:12.972 "base_bdevs_list": [ 00:13:12.972 { 00:13:12.972 "name": "BaseBdev1", 00:13:12.972 "uuid": "9dbe4582-9102-4dc3-abec-5e51365c79b4", 00:13:12.972 "is_configured": true, 00:13:12.972 "data_offset": 2048, 00:13:12.972 "data_size": 63488 00:13:12.972 }, 00:13:12.972 { 00:13:12.972 "name": "BaseBdev2", 00:13:12.972 "uuid": "81f8add4-c768-4ffb-9430-da8cf6fcf2a5", 00:13:12.972 "is_configured": true, 00:13:12.972 "data_offset": 2048, 00:13:12.972 "data_size": 63488 00:13:12.972 }, 00:13:12.972 { 00:13:12.972 "name": "BaseBdev3", 00:13:12.972 "uuid": "fb1d39bb-0e6e-485f-aabf-9ece41af1f40", 00:13:12.972 "is_configured": true, 00:13:12.972 "data_offset": 2048, 00:13:12.972 "data_size": 63488 00:13:12.972 } 00:13:12.972 ] 00:13:12.972 } 00:13:12.972 } 00:13:12.972 }' 00:13:12.972 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:13.241 BaseBdev2 00:13:13.241 BaseBdev3' 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.241 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.242 [2024-12-07 05:40:46.549655] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.242 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.541 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:13.541 "name": "Existed_Raid", 00:13:13.541 "uuid": "8ce50fc8-8335-42ba-833f-27b47b083a20", 00:13:13.541 "strip_size_kb": 64, 00:13:13.541 "state": "online", 00:13:13.541 "raid_level": "raid5f", 00:13:13.541 "superblock": true, 00:13:13.541 "num_base_bdevs": 3, 00:13:13.541 "num_base_bdevs_discovered": 2, 00:13:13.541 "num_base_bdevs_operational": 2, 00:13:13.541 "base_bdevs_list": [ 00:13:13.541 { 00:13:13.541 "name": null, 00:13:13.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.541 "is_configured": false, 00:13:13.541 "data_offset": 0, 00:13:13.541 "data_size": 63488 00:13:13.541 }, 00:13:13.541 { 00:13:13.541 "name": "BaseBdev2", 00:13:13.541 "uuid": "81f8add4-c768-4ffb-9430-da8cf6fcf2a5", 00:13:13.541 "is_configured": true, 00:13:13.541 "data_offset": 2048, 00:13:13.541 "data_size": 63488 00:13:13.541 }, 00:13:13.541 { 00:13:13.541 "name": "BaseBdev3", 00:13:13.541 "uuid": "fb1d39bb-0e6e-485f-aabf-9ece41af1f40", 00:13:13.541 "is_configured": true, 00:13:13.541 "data_offset": 2048, 00:13:13.541 "data_size": 63488 00:13:13.541 } 00:13:13.541 ] 00:13:13.541 }' 00:13:13.541 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:13.541 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.800 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:13.800 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:13.800 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:13.800 05:40:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.800 05:40:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.800 [2024-12-07 05:40:47.044125] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:13.800 [2024-12-07 05:40:47.044325] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:13.800 [2024-12-07 05:40:47.055494] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:13.800 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.801 [2024-12-07 05:40:47.099439] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:13.801 [2024-12-07 05:40:47.099488] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.801 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.061 BaseBdev2 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.061 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.061 [ 00:13:14.061 { 00:13:14.061 "name": "BaseBdev2", 00:13:14.061 "aliases": [ 00:13:14.061 "97291f20-783a-4c20-a11d-c37cb87f06b8" 00:13:14.061 ], 00:13:14.061 "product_name": "Malloc disk", 00:13:14.061 "block_size": 512, 00:13:14.061 "num_blocks": 65536, 00:13:14.061 "uuid": "97291f20-783a-4c20-a11d-c37cb87f06b8", 00:13:14.061 "assigned_rate_limits": { 00:13:14.061 "rw_ios_per_sec": 0, 00:13:14.061 "rw_mbytes_per_sec": 0, 00:13:14.061 "r_mbytes_per_sec": 0, 00:13:14.061 "w_mbytes_per_sec": 0 00:13:14.061 }, 00:13:14.061 "claimed": false, 00:13:14.061 "zoned": false, 00:13:14.061 "supported_io_types": { 00:13:14.061 "read": true, 00:13:14.061 "write": true, 00:13:14.061 "unmap": true, 00:13:14.061 "flush": true, 00:13:14.061 "reset": true, 00:13:14.061 "nvme_admin": false, 00:13:14.061 "nvme_io": false, 00:13:14.061 "nvme_io_md": false, 00:13:14.061 "write_zeroes": true, 00:13:14.061 "zcopy": true, 00:13:14.061 "get_zone_info": false, 00:13:14.061 "zone_management": false, 00:13:14.061 "zone_append": false, 00:13:14.061 "compare": false, 00:13:14.061 "compare_and_write": false, 00:13:14.061 "abort": true, 00:13:14.061 "seek_hole": false, 00:13:14.061 "seek_data": false, 00:13:14.061 "copy": true, 00:13:14.061 "nvme_iov_md": false 00:13:14.061 }, 00:13:14.061 "memory_domains": [ 00:13:14.061 { 00:13:14.061 "dma_device_id": "system", 00:13:14.061 "dma_device_type": 1 00:13:14.061 }, 00:13:14.061 { 00:13:14.061 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:14.061 "dma_device_type": 2 00:13:14.061 } 00:13:14.061 ], 00:13:14.061 "driver_specific": {} 00:13:14.061 } 00:13:14.061 ] 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.062 BaseBdev3 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.062 [ 00:13:14.062 { 00:13:14.062 "name": "BaseBdev3", 00:13:14.062 "aliases": [ 00:13:14.062 "38d27d75-f983-4214-ba0d-fa71bb3de8d6" 00:13:14.062 ], 00:13:14.062 "product_name": "Malloc disk", 00:13:14.062 "block_size": 512, 00:13:14.062 "num_blocks": 65536, 00:13:14.062 "uuid": "38d27d75-f983-4214-ba0d-fa71bb3de8d6", 00:13:14.062 "assigned_rate_limits": { 00:13:14.062 "rw_ios_per_sec": 0, 00:13:14.062 "rw_mbytes_per_sec": 0, 00:13:14.062 "r_mbytes_per_sec": 0, 00:13:14.062 "w_mbytes_per_sec": 0 00:13:14.062 }, 00:13:14.062 "claimed": false, 00:13:14.062 "zoned": false, 00:13:14.062 "supported_io_types": { 00:13:14.062 "read": true, 00:13:14.062 "write": true, 00:13:14.062 "unmap": true, 00:13:14.062 "flush": true, 00:13:14.062 "reset": true, 00:13:14.062 "nvme_admin": false, 00:13:14.062 "nvme_io": false, 00:13:14.062 "nvme_io_md": false, 00:13:14.062 "write_zeroes": true, 00:13:14.062 "zcopy": true, 00:13:14.062 "get_zone_info": false, 00:13:14.062 "zone_management": false, 00:13:14.062 "zone_append": false, 00:13:14.062 "compare": false, 00:13:14.062 "compare_and_write": false, 00:13:14.062 "abort": true, 00:13:14.062 "seek_hole": false, 00:13:14.062 "seek_data": false, 00:13:14.062 "copy": true, 00:13:14.062 "nvme_iov_md": false 00:13:14.062 }, 00:13:14.062 "memory_domains": [ 00:13:14.062 { 00:13:14.062 "dma_device_id": "system", 00:13:14.062 "dma_device_type": 1 00:13:14.062 }, 00:13:14.062 { 00:13:14.062 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:14.062 "dma_device_type": 2 00:13:14.062 } 00:13:14.062 ], 00:13:14.062 "driver_specific": {} 00:13:14.062 } 00:13:14.062 ] 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.062 [2024-12-07 05:40:47.266620] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:14.062 [2024-12-07 05:40:47.266717] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:14.062 [2024-12-07 05:40:47.266761] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:14.062 [2024-12-07 05:40:47.268534] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:14.062 "name": "Existed_Raid", 00:13:14.062 "uuid": "72e273c1-6161-47f2-be6d-afe1a9789831", 00:13:14.062 "strip_size_kb": 64, 00:13:14.062 "state": "configuring", 00:13:14.062 "raid_level": "raid5f", 00:13:14.062 "superblock": true, 00:13:14.062 "num_base_bdevs": 3, 00:13:14.062 "num_base_bdevs_discovered": 2, 00:13:14.062 "num_base_bdevs_operational": 3, 00:13:14.062 "base_bdevs_list": [ 00:13:14.062 { 00:13:14.062 "name": "BaseBdev1", 00:13:14.062 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.062 "is_configured": false, 00:13:14.062 "data_offset": 0, 00:13:14.062 "data_size": 0 00:13:14.062 }, 00:13:14.062 { 00:13:14.062 "name": "BaseBdev2", 00:13:14.062 "uuid": "97291f20-783a-4c20-a11d-c37cb87f06b8", 00:13:14.062 "is_configured": true, 00:13:14.062 "data_offset": 2048, 00:13:14.062 "data_size": 63488 00:13:14.062 }, 00:13:14.062 { 00:13:14.062 "name": "BaseBdev3", 00:13:14.062 "uuid": "38d27d75-f983-4214-ba0d-fa71bb3de8d6", 00:13:14.062 "is_configured": true, 00:13:14.062 "data_offset": 2048, 00:13:14.062 "data_size": 63488 00:13:14.062 } 00:13:14.062 ] 00:13:14.062 }' 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:14.062 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.633 [2024-12-07 05:40:47.717853] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:14.633 "name": "Existed_Raid", 00:13:14.633 "uuid": "72e273c1-6161-47f2-be6d-afe1a9789831", 00:13:14.633 "strip_size_kb": 64, 00:13:14.633 "state": "configuring", 00:13:14.633 "raid_level": "raid5f", 00:13:14.633 "superblock": true, 00:13:14.633 "num_base_bdevs": 3, 00:13:14.633 "num_base_bdevs_discovered": 1, 00:13:14.633 "num_base_bdevs_operational": 3, 00:13:14.633 "base_bdevs_list": [ 00:13:14.633 { 00:13:14.633 "name": "BaseBdev1", 00:13:14.633 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.633 "is_configured": false, 00:13:14.633 "data_offset": 0, 00:13:14.633 "data_size": 0 00:13:14.633 }, 00:13:14.633 { 00:13:14.633 "name": null, 00:13:14.633 "uuid": "97291f20-783a-4c20-a11d-c37cb87f06b8", 00:13:14.633 "is_configured": false, 00:13:14.633 "data_offset": 0, 00:13:14.633 "data_size": 63488 00:13:14.633 }, 00:13:14.633 { 00:13:14.633 "name": "BaseBdev3", 00:13:14.633 "uuid": "38d27d75-f983-4214-ba0d-fa71bb3de8d6", 00:13:14.633 "is_configured": true, 00:13:14.633 "data_offset": 2048, 00:13:14.633 "data_size": 63488 00:13:14.633 } 00:13:14.633 ] 00:13:14.633 }' 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:14.633 05:40:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.894 [2024-12-07 05:40:48.211959] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:14.894 BaseBdev1 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.894 [ 00:13:14.894 { 00:13:14.894 "name": "BaseBdev1", 00:13:14.894 "aliases": [ 00:13:14.894 "aa0ab300-4d9b-45b6-830b-19ff65e68fb4" 00:13:14.894 ], 00:13:14.894 "product_name": "Malloc disk", 00:13:14.894 "block_size": 512, 00:13:14.894 "num_blocks": 65536, 00:13:14.894 "uuid": "aa0ab300-4d9b-45b6-830b-19ff65e68fb4", 00:13:14.894 "assigned_rate_limits": { 00:13:14.894 "rw_ios_per_sec": 0, 00:13:14.894 "rw_mbytes_per_sec": 0, 00:13:14.894 "r_mbytes_per_sec": 0, 00:13:14.894 "w_mbytes_per_sec": 0 00:13:14.894 }, 00:13:14.894 "claimed": true, 00:13:14.894 "claim_type": "exclusive_write", 00:13:14.894 "zoned": false, 00:13:14.894 "supported_io_types": { 00:13:14.894 "read": true, 00:13:14.894 "write": true, 00:13:14.894 "unmap": true, 00:13:14.894 "flush": true, 00:13:14.894 "reset": true, 00:13:14.894 "nvme_admin": false, 00:13:14.894 "nvme_io": false, 00:13:14.894 "nvme_io_md": false, 00:13:14.894 "write_zeroes": true, 00:13:14.894 "zcopy": true, 00:13:14.894 "get_zone_info": false, 00:13:14.894 "zone_management": false, 00:13:14.894 "zone_append": false, 00:13:14.894 "compare": false, 00:13:14.894 "compare_and_write": false, 00:13:14.894 "abort": true, 00:13:14.894 "seek_hole": false, 00:13:14.894 "seek_data": false, 00:13:14.894 "copy": true, 00:13:14.894 "nvme_iov_md": false 00:13:14.894 }, 00:13:14.894 "memory_domains": [ 00:13:14.894 { 00:13:14.894 "dma_device_id": "system", 00:13:14.894 "dma_device_type": 1 00:13:14.894 }, 00:13:14.894 { 00:13:14.894 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:14.894 "dma_device_type": 2 00:13:14.894 } 00:13:14.894 ], 00:13:14.894 "driver_specific": {} 00:13:14.894 } 00:13:14.894 ] 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:14.894 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:14.895 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:15.155 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:15.155 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:15.155 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.155 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.155 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.155 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:15.155 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.155 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:15.155 "name": "Existed_Raid", 00:13:15.155 "uuid": "72e273c1-6161-47f2-be6d-afe1a9789831", 00:13:15.155 "strip_size_kb": 64, 00:13:15.155 "state": "configuring", 00:13:15.155 "raid_level": "raid5f", 00:13:15.155 "superblock": true, 00:13:15.155 "num_base_bdevs": 3, 00:13:15.155 "num_base_bdevs_discovered": 2, 00:13:15.155 "num_base_bdevs_operational": 3, 00:13:15.155 "base_bdevs_list": [ 00:13:15.155 { 00:13:15.155 "name": "BaseBdev1", 00:13:15.155 "uuid": "aa0ab300-4d9b-45b6-830b-19ff65e68fb4", 00:13:15.155 "is_configured": true, 00:13:15.155 "data_offset": 2048, 00:13:15.155 "data_size": 63488 00:13:15.155 }, 00:13:15.155 { 00:13:15.155 "name": null, 00:13:15.155 "uuid": "97291f20-783a-4c20-a11d-c37cb87f06b8", 00:13:15.155 "is_configured": false, 00:13:15.155 "data_offset": 0, 00:13:15.155 "data_size": 63488 00:13:15.155 }, 00:13:15.155 { 00:13:15.155 "name": "BaseBdev3", 00:13:15.155 "uuid": "38d27d75-f983-4214-ba0d-fa71bb3de8d6", 00:13:15.155 "is_configured": true, 00:13:15.155 "data_offset": 2048, 00:13:15.155 "data_size": 63488 00:13:15.155 } 00:13:15.155 ] 00:13:15.155 }' 00:13:15.155 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:15.155 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.415 [2024-12-07 05:40:48.671228] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.415 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:15.415 "name": "Existed_Raid", 00:13:15.415 "uuid": "72e273c1-6161-47f2-be6d-afe1a9789831", 00:13:15.415 "strip_size_kb": 64, 00:13:15.415 "state": "configuring", 00:13:15.415 "raid_level": "raid5f", 00:13:15.415 "superblock": true, 00:13:15.415 "num_base_bdevs": 3, 00:13:15.415 "num_base_bdevs_discovered": 1, 00:13:15.415 "num_base_bdevs_operational": 3, 00:13:15.415 "base_bdevs_list": [ 00:13:15.415 { 00:13:15.415 "name": "BaseBdev1", 00:13:15.415 "uuid": "aa0ab300-4d9b-45b6-830b-19ff65e68fb4", 00:13:15.415 "is_configured": true, 00:13:15.415 "data_offset": 2048, 00:13:15.415 "data_size": 63488 00:13:15.415 }, 00:13:15.415 { 00:13:15.415 "name": null, 00:13:15.415 "uuid": "97291f20-783a-4c20-a11d-c37cb87f06b8", 00:13:15.415 "is_configured": false, 00:13:15.415 "data_offset": 0, 00:13:15.415 "data_size": 63488 00:13:15.415 }, 00:13:15.415 { 00:13:15.415 "name": null, 00:13:15.415 "uuid": "38d27d75-f983-4214-ba0d-fa71bb3de8d6", 00:13:15.415 "is_configured": false, 00:13:15.415 "data_offset": 0, 00:13:15.415 "data_size": 63488 00:13:15.415 } 00:13:15.416 ] 00:13:15.416 }' 00:13:15.416 05:40:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:15.416 05:40:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.788 [2024-12-07 05:40:49.114494] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:15.788 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.048 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:16.048 "name": "Existed_Raid", 00:13:16.048 "uuid": "72e273c1-6161-47f2-be6d-afe1a9789831", 00:13:16.048 "strip_size_kb": 64, 00:13:16.048 "state": "configuring", 00:13:16.048 "raid_level": "raid5f", 00:13:16.048 "superblock": true, 00:13:16.048 "num_base_bdevs": 3, 00:13:16.048 "num_base_bdevs_discovered": 2, 00:13:16.048 "num_base_bdevs_operational": 3, 00:13:16.048 "base_bdevs_list": [ 00:13:16.048 { 00:13:16.048 "name": "BaseBdev1", 00:13:16.048 "uuid": "aa0ab300-4d9b-45b6-830b-19ff65e68fb4", 00:13:16.048 "is_configured": true, 00:13:16.048 "data_offset": 2048, 00:13:16.048 "data_size": 63488 00:13:16.048 }, 00:13:16.048 { 00:13:16.048 "name": null, 00:13:16.048 "uuid": "97291f20-783a-4c20-a11d-c37cb87f06b8", 00:13:16.048 "is_configured": false, 00:13:16.048 "data_offset": 0, 00:13:16.048 "data_size": 63488 00:13:16.048 }, 00:13:16.048 { 00:13:16.048 "name": "BaseBdev3", 00:13:16.048 "uuid": "38d27d75-f983-4214-ba0d-fa71bb3de8d6", 00:13:16.048 "is_configured": true, 00:13:16.048 "data_offset": 2048, 00:13:16.048 "data_size": 63488 00:13:16.048 } 00:13:16.048 ] 00:13:16.048 }' 00:13:16.048 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:16.048 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.308 [2024-12-07 05:40:49.609679] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:16.308 "name": "Existed_Raid", 00:13:16.308 "uuid": "72e273c1-6161-47f2-be6d-afe1a9789831", 00:13:16.308 "strip_size_kb": 64, 00:13:16.308 "state": "configuring", 00:13:16.308 "raid_level": "raid5f", 00:13:16.308 "superblock": true, 00:13:16.308 "num_base_bdevs": 3, 00:13:16.308 "num_base_bdevs_discovered": 1, 00:13:16.308 "num_base_bdevs_operational": 3, 00:13:16.308 "base_bdevs_list": [ 00:13:16.308 { 00:13:16.308 "name": null, 00:13:16.308 "uuid": "aa0ab300-4d9b-45b6-830b-19ff65e68fb4", 00:13:16.308 "is_configured": false, 00:13:16.308 "data_offset": 0, 00:13:16.308 "data_size": 63488 00:13:16.308 }, 00:13:16.308 { 00:13:16.308 "name": null, 00:13:16.308 "uuid": "97291f20-783a-4c20-a11d-c37cb87f06b8", 00:13:16.308 "is_configured": false, 00:13:16.308 "data_offset": 0, 00:13:16.308 "data_size": 63488 00:13:16.308 }, 00:13:16.308 { 00:13:16.308 "name": "BaseBdev3", 00:13:16.308 "uuid": "38d27d75-f983-4214-ba0d-fa71bb3de8d6", 00:13:16.308 "is_configured": true, 00:13:16.308 "data_offset": 2048, 00:13:16.308 "data_size": 63488 00:13:16.308 } 00:13:16.308 ] 00:13:16.308 }' 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:16.308 05:40:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.879 [2024-12-07 05:40:50.127302] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:16.879 "name": "Existed_Raid", 00:13:16.879 "uuid": "72e273c1-6161-47f2-be6d-afe1a9789831", 00:13:16.879 "strip_size_kb": 64, 00:13:16.879 "state": "configuring", 00:13:16.879 "raid_level": "raid5f", 00:13:16.879 "superblock": true, 00:13:16.879 "num_base_bdevs": 3, 00:13:16.879 "num_base_bdevs_discovered": 2, 00:13:16.879 "num_base_bdevs_operational": 3, 00:13:16.879 "base_bdevs_list": [ 00:13:16.879 { 00:13:16.879 "name": null, 00:13:16.879 "uuid": "aa0ab300-4d9b-45b6-830b-19ff65e68fb4", 00:13:16.879 "is_configured": false, 00:13:16.879 "data_offset": 0, 00:13:16.879 "data_size": 63488 00:13:16.879 }, 00:13:16.879 { 00:13:16.879 "name": "BaseBdev2", 00:13:16.879 "uuid": "97291f20-783a-4c20-a11d-c37cb87f06b8", 00:13:16.879 "is_configured": true, 00:13:16.879 "data_offset": 2048, 00:13:16.879 "data_size": 63488 00:13:16.879 }, 00:13:16.879 { 00:13:16.879 "name": "BaseBdev3", 00:13:16.879 "uuid": "38d27d75-f983-4214-ba0d-fa71bb3de8d6", 00:13:16.879 "is_configured": true, 00:13:16.879 "data_offset": 2048, 00:13:16.879 "data_size": 63488 00:13:16.879 } 00:13:16.879 ] 00:13:16.879 }' 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:16.879 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u aa0ab300-4d9b-45b6-830b-19ff65e68fb4 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.449 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.449 [2024-12-07 05:40:50.649394] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:17.449 [2024-12-07 05:40:50.649650] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:17.449 [2024-12-07 05:40:50.649714] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:17.449 NewBaseBdev 00:13:17.449 [2024-12-07 05:40:50.649969] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:17.449 [2024-12-07 05:40:50.650363] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:17.449 [2024-12-07 05:40:50.650424] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:13:17.449 [2024-12-07 05:40:50.650577] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.450 [ 00:13:17.450 { 00:13:17.450 "name": "NewBaseBdev", 00:13:17.450 "aliases": [ 00:13:17.450 "aa0ab300-4d9b-45b6-830b-19ff65e68fb4" 00:13:17.450 ], 00:13:17.450 "product_name": "Malloc disk", 00:13:17.450 "block_size": 512, 00:13:17.450 "num_blocks": 65536, 00:13:17.450 "uuid": "aa0ab300-4d9b-45b6-830b-19ff65e68fb4", 00:13:17.450 "assigned_rate_limits": { 00:13:17.450 "rw_ios_per_sec": 0, 00:13:17.450 "rw_mbytes_per_sec": 0, 00:13:17.450 "r_mbytes_per_sec": 0, 00:13:17.450 "w_mbytes_per_sec": 0 00:13:17.450 }, 00:13:17.450 "claimed": true, 00:13:17.450 "claim_type": "exclusive_write", 00:13:17.450 "zoned": false, 00:13:17.450 "supported_io_types": { 00:13:17.450 "read": true, 00:13:17.450 "write": true, 00:13:17.450 "unmap": true, 00:13:17.450 "flush": true, 00:13:17.450 "reset": true, 00:13:17.450 "nvme_admin": false, 00:13:17.450 "nvme_io": false, 00:13:17.450 "nvme_io_md": false, 00:13:17.450 "write_zeroes": true, 00:13:17.450 "zcopy": true, 00:13:17.450 "get_zone_info": false, 00:13:17.450 "zone_management": false, 00:13:17.450 "zone_append": false, 00:13:17.450 "compare": false, 00:13:17.450 "compare_and_write": false, 00:13:17.450 "abort": true, 00:13:17.450 "seek_hole": false, 00:13:17.450 "seek_data": false, 00:13:17.450 "copy": true, 00:13:17.450 "nvme_iov_md": false 00:13:17.450 }, 00:13:17.450 "memory_domains": [ 00:13:17.450 { 00:13:17.450 "dma_device_id": "system", 00:13:17.450 "dma_device_type": 1 00:13:17.450 }, 00:13:17.450 { 00:13:17.450 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:17.450 "dma_device_type": 2 00:13:17.450 } 00:13:17.450 ], 00:13:17.450 "driver_specific": {} 00:13:17.450 } 00:13:17.450 ] 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:17.450 "name": "Existed_Raid", 00:13:17.450 "uuid": "72e273c1-6161-47f2-be6d-afe1a9789831", 00:13:17.450 "strip_size_kb": 64, 00:13:17.450 "state": "online", 00:13:17.450 "raid_level": "raid5f", 00:13:17.450 "superblock": true, 00:13:17.450 "num_base_bdevs": 3, 00:13:17.450 "num_base_bdevs_discovered": 3, 00:13:17.450 "num_base_bdevs_operational": 3, 00:13:17.450 "base_bdevs_list": [ 00:13:17.450 { 00:13:17.450 "name": "NewBaseBdev", 00:13:17.450 "uuid": "aa0ab300-4d9b-45b6-830b-19ff65e68fb4", 00:13:17.450 "is_configured": true, 00:13:17.450 "data_offset": 2048, 00:13:17.450 "data_size": 63488 00:13:17.450 }, 00:13:17.450 { 00:13:17.450 "name": "BaseBdev2", 00:13:17.450 "uuid": "97291f20-783a-4c20-a11d-c37cb87f06b8", 00:13:17.450 "is_configured": true, 00:13:17.450 "data_offset": 2048, 00:13:17.450 "data_size": 63488 00:13:17.450 }, 00:13:17.450 { 00:13:17.450 "name": "BaseBdev3", 00:13:17.450 "uuid": "38d27d75-f983-4214-ba0d-fa71bb3de8d6", 00:13:17.450 "is_configured": true, 00:13:17.450 "data_offset": 2048, 00:13:17.450 "data_size": 63488 00:13:17.450 } 00:13:17.450 ] 00:13:17.450 }' 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:17.450 05:40:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.020 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:18.020 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:18.020 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:18.020 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:18.020 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:18.021 [2024-12-07 05:40:51.120806] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:18.021 "name": "Existed_Raid", 00:13:18.021 "aliases": [ 00:13:18.021 "72e273c1-6161-47f2-be6d-afe1a9789831" 00:13:18.021 ], 00:13:18.021 "product_name": "Raid Volume", 00:13:18.021 "block_size": 512, 00:13:18.021 "num_blocks": 126976, 00:13:18.021 "uuid": "72e273c1-6161-47f2-be6d-afe1a9789831", 00:13:18.021 "assigned_rate_limits": { 00:13:18.021 "rw_ios_per_sec": 0, 00:13:18.021 "rw_mbytes_per_sec": 0, 00:13:18.021 "r_mbytes_per_sec": 0, 00:13:18.021 "w_mbytes_per_sec": 0 00:13:18.021 }, 00:13:18.021 "claimed": false, 00:13:18.021 "zoned": false, 00:13:18.021 "supported_io_types": { 00:13:18.021 "read": true, 00:13:18.021 "write": true, 00:13:18.021 "unmap": false, 00:13:18.021 "flush": false, 00:13:18.021 "reset": true, 00:13:18.021 "nvme_admin": false, 00:13:18.021 "nvme_io": false, 00:13:18.021 "nvme_io_md": false, 00:13:18.021 "write_zeroes": true, 00:13:18.021 "zcopy": false, 00:13:18.021 "get_zone_info": false, 00:13:18.021 "zone_management": false, 00:13:18.021 "zone_append": false, 00:13:18.021 "compare": false, 00:13:18.021 "compare_and_write": false, 00:13:18.021 "abort": false, 00:13:18.021 "seek_hole": false, 00:13:18.021 "seek_data": false, 00:13:18.021 "copy": false, 00:13:18.021 "nvme_iov_md": false 00:13:18.021 }, 00:13:18.021 "driver_specific": { 00:13:18.021 "raid": { 00:13:18.021 "uuid": "72e273c1-6161-47f2-be6d-afe1a9789831", 00:13:18.021 "strip_size_kb": 64, 00:13:18.021 "state": "online", 00:13:18.021 "raid_level": "raid5f", 00:13:18.021 "superblock": true, 00:13:18.021 "num_base_bdevs": 3, 00:13:18.021 "num_base_bdevs_discovered": 3, 00:13:18.021 "num_base_bdevs_operational": 3, 00:13:18.021 "base_bdevs_list": [ 00:13:18.021 { 00:13:18.021 "name": "NewBaseBdev", 00:13:18.021 "uuid": "aa0ab300-4d9b-45b6-830b-19ff65e68fb4", 00:13:18.021 "is_configured": true, 00:13:18.021 "data_offset": 2048, 00:13:18.021 "data_size": 63488 00:13:18.021 }, 00:13:18.021 { 00:13:18.021 "name": "BaseBdev2", 00:13:18.021 "uuid": "97291f20-783a-4c20-a11d-c37cb87f06b8", 00:13:18.021 "is_configured": true, 00:13:18.021 "data_offset": 2048, 00:13:18.021 "data_size": 63488 00:13:18.021 }, 00:13:18.021 { 00:13:18.021 "name": "BaseBdev3", 00:13:18.021 "uuid": "38d27d75-f983-4214-ba0d-fa71bb3de8d6", 00:13:18.021 "is_configured": true, 00:13:18.021 "data_offset": 2048, 00:13:18.021 "data_size": 63488 00:13:18.021 } 00:13:18.021 ] 00:13:18.021 } 00:13:18.021 } 00:13:18.021 }' 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:18.021 BaseBdev2 00:13:18.021 BaseBdev3' 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.021 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.021 [2024-12-07 05:40:51.372165] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:18.021 [2024-12-07 05:40:51.372189] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:18.021 [2024-12-07 05:40:51.372260] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:18.021 [2024-12-07 05:40:51.372492] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:18.021 [2024-12-07 05:40:51.372511] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:13:18.022 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.022 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 90670 00:13:18.022 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 90670 ']' 00:13:18.022 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 90670 00:13:18.022 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:13:18.022 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:18.281 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 90670 00:13:18.281 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:18.281 killing process with pid 90670 00:13:18.281 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:18.281 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 90670' 00:13:18.281 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 90670 00:13:18.281 [2024-12-07 05:40:51.418876] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:18.281 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 90670 00:13:18.281 [2024-12-07 05:40:51.449248] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:18.541 05:40:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:13:18.541 00:13:18.541 real 0m8.602s 00:13:18.541 user 0m14.680s 00:13:18.541 sys 0m1.674s 00:13:18.541 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:18.541 05:40:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.541 ************************************ 00:13:18.541 END TEST raid5f_state_function_test_sb 00:13:18.541 ************************************ 00:13:18.542 05:40:51 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:13:18.542 05:40:51 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:13:18.542 05:40:51 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:18.542 05:40:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:18.542 ************************************ 00:13:18.542 START TEST raid5f_superblock_test 00:13:18.542 ************************************ 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 3 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=91268 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 91268 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 91268 ']' 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:18.542 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:18.542 05:40:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:18.542 [2024-12-07 05:40:51.819911] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:13:18.542 [2024-12-07 05:40:51.820113] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91268 ] 00:13:18.802 [2024-12-07 05:40:51.952373] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:18.802 [2024-12-07 05:40:51.976808] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:18.802 [2024-12-07 05:40:52.019234] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:18.802 [2024-12-07 05:40:52.019358] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.372 malloc1 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.372 [2024-12-07 05:40:52.674736] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:19.372 [2024-12-07 05:40:52.674792] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:19.372 [2024-12-07 05:40:52.674829] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:19.372 [2024-12-07 05:40:52.674842] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:19.372 [2024-12-07 05:40:52.676865] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:19.372 [2024-12-07 05:40:52.676945] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:19.372 pt1 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.372 malloc2 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.372 [2024-12-07 05:40:52.703291] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:19.372 [2024-12-07 05:40:52.703364] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:19.372 [2024-12-07 05:40:52.703383] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:19.372 [2024-12-07 05:40:52.703392] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:19.372 [2024-12-07 05:40:52.705511] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:19.372 [2024-12-07 05:40:52.705550] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:19.372 pt2 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.372 malloc3 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.372 [2024-12-07 05:40:52.731990] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:19.372 [2024-12-07 05:40:52.732087] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:19.372 [2024-12-07 05:40:52.732122] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:19.372 [2024-12-07 05:40:52.732152] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:19.372 [2024-12-07 05:40:52.734280] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:19.372 [2024-12-07 05:40:52.734355] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:19.372 pt3 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.372 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:19.632 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:19.632 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:13:19.632 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.632 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.632 [2024-12-07 05:40:52.744039] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:19.633 [2024-12-07 05:40:52.745894] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:19.633 [2024-12-07 05:40:52.746003] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:19.633 [2024-12-07 05:40:52.746179] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:19.633 [2024-12-07 05:40:52.746226] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:19.633 [2024-12-07 05:40:52.746511] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:13:19.633 [2024-12-07 05:40:52.746971] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:19.633 [2024-12-07 05:40:52.747027] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:19.633 [2024-12-07 05:40:52.747174] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:19.633 "name": "raid_bdev1", 00:13:19.633 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:19.633 "strip_size_kb": 64, 00:13:19.633 "state": "online", 00:13:19.633 "raid_level": "raid5f", 00:13:19.633 "superblock": true, 00:13:19.633 "num_base_bdevs": 3, 00:13:19.633 "num_base_bdevs_discovered": 3, 00:13:19.633 "num_base_bdevs_operational": 3, 00:13:19.633 "base_bdevs_list": [ 00:13:19.633 { 00:13:19.633 "name": "pt1", 00:13:19.633 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:19.633 "is_configured": true, 00:13:19.633 "data_offset": 2048, 00:13:19.633 "data_size": 63488 00:13:19.633 }, 00:13:19.633 { 00:13:19.633 "name": "pt2", 00:13:19.633 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:19.633 "is_configured": true, 00:13:19.633 "data_offset": 2048, 00:13:19.633 "data_size": 63488 00:13:19.633 }, 00:13:19.633 { 00:13:19.633 "name": "pt3", 00:13:19.633 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:19.633 "is_configured": true, 00:13:19.633 "data_offset": 2048, 00:13:19.633 "data_size": 63488 00:13:19.633 } 00:13:19.633 ] 00:13:19.633 }' 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:19.633 05:40:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.893 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:13:19.893 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:19.893 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:19.893 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:19.893 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:19.893 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:19.893 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:19.893 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:19.893 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.893 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:19.893 [2024-12-07 05:40:53.223880] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:19.893 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.153 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:20.153 "name": "raid_bdev1", 00:13:20.153 "aliases": [ 00:13:20.153 "681a9622-6617-42d6-8318-020dbde3bdd8" 00:13:20.153 ], 00:13:20.153 "product_name": "Raid Volume", 00:13:20.153 "block_size": 512, 00:13:20.153 "num_blocks": 126976, 00:13:20.153 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:20.153 "assigned_rate_limits": { 00:13:20.153 "rw_ios_per_sec": 0, 00:13:20.153 "rw_mbytes_per_sec": 0, 00:13:20.153 "r_mbytes_per_sec": 0, 00:13:20.153 "w_mbytes_per_sec": 0 00:13:20.153 }, 00:13:20.153 "claimed": false, 00:13:20.153 "zoned": false, 00:13:20.153 "supported_io_types": { 00:13:20.153 "read": true, 00:13:20.153 "write": true, 00:13:20.153 "unmap": false, 00:13:20.153 "flush": false, 00:13:20.153 "reset": true, 00:13:20.153 "nvme_admin": false, 00:13:20.153 "nvme_io": false, 00:13:20.153 "nvme_io_md": false, 00:13:20.153 "write_zeroes": true, 00:13:20.153 "zcopy": false, 00:13:20.153 "get_zone_info": false, 00:13:20.153 "zone_management": false, 00:13:20.153 "zone_append": false, 00:13:20.153 "compare": false, 00:13:20.153 "compare_and_write": false, 00:13:20.153 "abort": false, 00:13:20.153 "seek_hole": false, 00:13:20.153 "seek_data": false, 00:13:20.153 "copy": false, 00:13:20.153 "nvme_iov_md": false 00:13:20.153 }, 00:13:20.153 "driver_specific": { 00:13:20.153 "raid": { 00:13:20.153 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:20.153 "strip_size_kb": 64, 00:13:20.153 "state": "online", 00:13:20.153 "raid_level": "raid5f", 00:13:20.153 "superblock": true, 00:13:20.153 "num_base_bdevs": 3, 00:13:20.153 "num_base_bdevs_discovered": 3, 00:13:20.153 "num_base_bdevs_operational": 3, 00:13:20.153 "base_bdevs_list": [ 00:13:20.153 { 00:13:20.153 "name": "pt1", 00:13:20.153 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:20.154 "is_configured": true, 00:13:20.154 "data_offset": 2048, 00:13:20.154 "data_size": 63488 00:13:20.154 }, 00:13:20.154 { 00:13:20.154 "name": "pt2", 00:13:20.154 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:20.154 "is_configured": true, 00:13:20.154 "data_offset": 2048, 00:13:20.154 "data_size": 63488 00:13:20.154 }, 00:13:20.154 { 00:13:20.154 "name": "pt3", 00:13:20.154 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:20.154 "is_configured": true, 00:13:20.154 "data_offset": 2048, 00:13:20.154 "data_size": 63488 00:13:20.154 } 00:13:20.154 ] 00:13:20.154 } 00:13:20.154 } 00:13:20.154 }' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:20.154 pt2 00:13:20.154 pt3' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.154 [2024-12-07 05:40:53.487373] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=681a9622-6617-42d6-8318-020dbde3bdd8 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 681a9622-6617-42d6-8318-020dbde3bdd8 ']' 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.154 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.154 [2024-12-07 05:40:53.515159] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:20.154 [2024-12-07 05:40:53.515181] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:20.154 [2024-12-07 05:40:53.515255] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:20.154 [2024-12-07 05:40:53.515329] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:20.154 [2024-12-07 05:40:53.515344] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.415 [2024-12-07 05:40:53.666927] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:13:20.415 [2024-12-07 05:40:53.668724] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:13:20.415 [2024-12-07 05:40:53.668817] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:13:20.415 [2024-12-07 05:40:53.668868] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:13:20.415 [2024-12-07 05:40:53.668917] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:13:20.415 [2024-12-07 05:40:53.668940] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:13:20.415 [2024-12-07 05:40:53.668951] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:20.415 [2024-12-07 05:40:53.668963] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:13:20.415 request: 00:13:20.415 { 00:13:20.415 "name": "raid_bdev1", 00:13:20.415 "raid_level": "raid5f", 00:13:20.415 "base_bdevs": [ 00:13:20.415 "malloc1", 00:13:20.415 "malloc2", 00:13:20.415 "malloc3" 00:13:20.415 ], 00:13:20.415 "strip_size_kb": 64, 00:13:20.415 "superblock": false, 00:13:20.415 "method": "bdev_raid_create", 00:13:20.415 "req_id": 1 00:13:20.415 } 00:13:20.415 Got JSON-RPC error response 00:13:20.415 response: 00:13:20.415 { 00:13:20.415 "code": -17, 00:13:20.415 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:13:20.415 } 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.415 [2024-12-07 05:40:53.726789] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:20.415 [2024-12-07 05:40:53.726876] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:20.415 [2024-12-07 05:40:53.726907] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:20.415 [2024-12-07 05:40:53.726938] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:20.415 [2024-12-07 05:40:53.728988] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:20.415 [2024-12-07 05:40:53.729061] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:20.415 [2024-12-07 05:40:53.729143] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:20.415 [2024-12-07 05:40:53.729207] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:20.415 pt1 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:20.415 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.416 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:20.416 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.416 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.416 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.416 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:20.416 "name": "raid_bdev1", 00:13:20.416 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:20.416 "strip_size_kb": 64, 00:13:20.416 "state": "configuring", 00:13:20.416 "raid_level": "raid5f", 00:13:20.416 "superblock": true, 00:13:20.416 "num_base_bdevs": 3, 00:13:20.416 "num_base_bdevs_discovered": 1, 00:13:20.416 "num_base_bdevs_operational": 3, 00:13:20.416 "base_bdevs_list": [ 00:13:20.416 { 00:13:20.416 "name": "pt1", 00:13:20.416 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:20.416 "is_configured": true, 00:13:20.416 "data_offset": 2048, 00:13:20.416 "data_size": 63488 00:13:20.416 }, 00:13:20.416 { 00:13:20.416 "name": null, 00:13:20.416 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:20.416 "is_configured": false, 00:13:20.416 "data_offset": 2048, 00:13:20.416 "data_size": 63488 00:13:20.416 }, 00:13:20.416 { 00:13:20.416 "name": null, 00:13:20.416 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:20.416 "is_configured": false, 00:13:20.416 "data_offset": 2048, 00:13:20.416 "data_size": 63488 00:13:20.416 } 00:13:20.416 ] 00:13:20.416 }' 00:13:20.416 05:40:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:20.416 05:40:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.985 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:13:20.985 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:20.985 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.985 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.985 [2024-12-07 05:40:54.134166] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:20.985 [2024-12-07 05:40:54.134230] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:20.985 [2024-12-07 05:40:54.134252] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:13:20.985 [2024-12-07 05:40:54.134265] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:20.985 [2024-12-07 05:40:54.134677] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:20.985 [2024-12-07 05:40:54.134700] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:20.986 [2024-12-07 05:40:54.134768] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:20.986 [2024-12-07 05:40:54.134791] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:20.986 pt2 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.986 [2024-12-07 05:40:54.146149] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:20.986 "name": "raid_bdev1", 00:13:20.986 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:20.986 "strip_size_kb": 64, 00:13:20.986 "state": "configuring", 00:13:20.986 "raid_level": "raid5f", 00:13:20.986 "superblock": true, 00:13:20.986 "num_base_bdevs": 3, 00:13:20.986 "num_base_bdevs_discovered": 1, 00:13:20.986 "num_base_bdevs_operational": 3, 00:13:20.986 "base_bdevs_list": [ 00:13:20.986 { 00:13:20.986 "name": "pt1", 00:13:20.986 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:20.986 "is_configured": true, 00:13:20.986 "data_offset": 2048, 00:13:20.986 "data_size": 63488 00:13:20.986 }, 00:13:20.986 { 00:13:20.986 "name": null, 00:13:20.986 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:20.986 "is_configured": false, 00:13:20.986 "data_offset": 0, 00:13:20.986 "data_size": 63488 00:13:20.986 }, 00:13:20.986 { 00:13:20.986 "name": null, 00:13:20.986 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:20.986 "is_configured": false, 00:13:20.986 "data_offset": 2048, 00:13:20.986 "data_size": 63488 00:13:20.986 } 00:13:20.986 ] 00:13:20.986 }' 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:20.986 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:21.246 [2024-12-07 05:40:54.585358] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:21.246 [2024-12-07 05:40:54.585452] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:21.246 [2024-12-07 05:40:54.585487] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:13:21.246 [2024-12-07 05:40:54.585514] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:21.246 [2024-12-07 05:40:54.585887] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:21.246 [2024-12-07 05:40:54.585945] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:21.246 [2024-12-07 05:40:54.586040] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:21.246 [2024-12-07 05:40:54.586088] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:21.246 pt2 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:21.246 [2024-12-07 05:40:54.597342] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:21.246 [2024-12-07 05:40:54.597384] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:21.246 [2024-12-07 05:40:54.597401] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:21.246 [2024-12-07 05:40:54.597409] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:21.246 [2024-12-07 05:40:54.597742] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:21.246 [2024-12-07 05:40:54.597760] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:21.246 [2024-12-07 05:40:54.597810] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:21.246 [2024-12-07 05:40:54.597826] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:21.246 [2024-12-07 05:40:54.597916] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:21.246 [2024-12-07 05:40:54.597939] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:21.246 [2024-12-07 05:40:54.598153] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:21.246 [2024-12-07 05:40:54.598523] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:21.246 [2024-12-07 05:40:54.598550] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:13:21.246 [2024-12-07 05:40:54.598665] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:21.246 pt3 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.246 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:21.247 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:21.506 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.506 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:21.506 "name": "raid_bdev1", 00:13:21.506 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:21.506 "strip_size_kb": 64, 00:13:21.506 "state": "online", 00:13:21.506 "raid_level": "raid5f", 00:13:21.506 "superblock": true, 00:13:21.506 "num_base_bdevs": 3, 00:13:21.506 "num_base_bdevs_discovered": 3, 00:13:21.506 "num_base_bdevs_operational": 3, 00:13:21.506 "base_bdevs_list": [ 00:13:21.506 { 00:13:21.506 "name": "pt1", 00:13:21.506 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:21.506 "is_configured": true, 00:13:21.506 "data_offset": 2048, 00:13:21.506 "data_size": 63488 00:13:21.506 }, 00:13:21.506 { 00:13:21.506 "name": "pt2", 00:13:21.506 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:21.506 "is_configured": true, 00:13:21.506 "data_offset": 2048, 00:13:21.506 "data_size": 63488 00:13:21.506 }, 00:13:21.506 { 00:13:21.506 "name": "pt3", 00:13:21.506 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:21.506 "is_configured": true, 00:13:21.506 "data_offset": 2048, 00:13:21.506 "data_size": 63488 00:13:21.506 } 00:13:21.506 ] 00:13:21.506 }' 00:13:21.506 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:21.506 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:21.766 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:13:21.766 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:21.766 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:21.766 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:21.766 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:21.766 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:21.766 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:21.766 05:40:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:21.766 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.766 05:40:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:21.766 [2024-12-07 05:40:54.992872] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:21.766 "name": "raid_bdev1", 00:13:21.766 "aliases": [ 00:13:21.766 "681a9622-6617-42d6-8318-020dbde3bdd8" 00:13:21.766 ], 00:13:21.766 "product_name": "Raid Volume", 00:13:21.766 "block_size": 512, 00:13:21.766 "num_blocks": 126976, 00:13:21.766 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:21.766 "assigned_rate_limits": { 00:13:21.766 "rw_ios_per_sec": 0, 00:13:21.766 "rw_mbytes_per_sec": 0, 00:13:21.766 "r_mbytes_per_sec": 0, 00:13:21.766 "w_mbytes_per_sec": 0 00:13:21.766 }, 00:13:21.766 "claimed": false, 00:13:21.766 "zoned": false, 00:13:21.766 "supported_io_types": { 00:13:21.766 "read": true, 00:13:21.766 "write": true, 00:13:21.766 "unmap": false, 00:13:21.766 "flush": false, 00:13:21.766 "reset": true, 00:13:21.766 "nvme_admin": false, 00:13:21.766 "nvme_io": false, 00:13:21.766 "nvme_io_md": false, 00:13:21.766 "write_zeroes": true, 00:13:21.766 "zcopy": false, 00:13:21.766 "get_zone_info": false, 00:13:21.766 "zone_management": false, 00:13:21.766 "zone_append": false, 00:13:21.766 "compare": false, 00:13:21.766 "compare_and_write": false, 00:13:21.766 "abort": false, 00:13:21.766 "seek_hole": false, 00:13:21.766 "seek_data": false, 00:13:21.766 "copy": false, 00:13:21.766 "nvme_iov_md": false 00:13:21.766 }, 00:13:21.766 "driver_specific": { 00:13:21.766 "raid": { 00:13:21.766 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:21.766 "strip_size_kb": 64, 00:13:21.766 "state": "online", 00:13:21.766 "raid_level": "raid5f", 00:13:21.766 "superblock": true, 00:13:21.766 "num_base_bdevs": 3, 00:13:21.766 "num_base_bdevs_discovered": 3, 00:13:21.766 "num_base_bdevs_operational": 3, 00:13:21.766 "base_bdevs_list": [ 00:13:21.766 { 00:13:21.766 "name": "pt1", 00:13:21.766 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:21.766 "is_configured": true, 00:13:21.766 "data_offset": 2048, 00:13:21.766 "data_size": 63488 00:13:21.766 }, 00:13:21.766 { 00:13:21.766 "name": "pt2", 00:13:21.766 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:21.766 "is_configured": true, 00:13:21.766 "data_offset": 2048, 00:13:21.766 "data_size": 63488 00:13:21.766 }, 00:13:21.766 { 00:13:21.766 "name": "pt3", 00:13:21.766 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:21.766 "is_configured": true, 00:13:21.766 "data_offset": 2048, 00:13:21.766 "data_size": 63488 00:13:21.766 } 00:13:21.766 ] 00:13:21.766 } 00:13:21.766 } 00:13:21.766 }' 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:21.766 pt2 00:13:21.766 pt3' 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:21.766 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.026 [2024-12-07 05:40:55.228392] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 681a9622-6617-42d6-8318-020dbde3bdd8 '!=' 681a9622-6617-42d6-8318-020dbde3bdd8 ']' 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.026 [2024-12-07 05:40:55.276193] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:22.026 "name": "raid_bdev1", 00:13:22.026 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:22.026 "strip_size_kb": 64, 00:13:22.026 "state": "online", 00:13:22.026 "raid_level": "raid5f", 00:13:22.026 "superblock": true, 00:13:22.026 "num_base_bdevs": 3, 00:13:22.026 "num_base_bdevs_discovered": 2, 00:13:22.026 "num_base_bdevs_operational": 2, 00:13:22.026 "base_bdevs_list": [ 00:13:22.026 { 00:13:22.026 "name": null, 00:13:22.026 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:22.026 "is_configured": false, 00:13:22.026 "data_offset": 0, 00:13:22.026 "data_size": 63488 00:13:22.026 }, 00:13:22.026 { 00:13:22.026 "name": "pt2", 00:13:22.026 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:22.026 "is_configured": true, 00:13:22.026 "data_offset": 2048, 00:13:22.026 "data_size": 63488 00:13:22.026 }, 00:13:22.026 { 00:13:22.026 "name": "pt3", 00:13:22.026 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:22.026 "is_configured": true, 00:13:22.026 "data_offset": 2048, 00:13:22.026 "data_size": 63488 00:13:22.026 } 00:13:22.026 ] 00:13:22.026 }' 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:22.026 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.594 [2024-12-07 05:40:55.699444] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:22.594 [2024-12-07 05:40:55.699523] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:22.594 [2024-12-07 05:40:55.699598] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:22.594 [2024-12-07 05:40:55.699681] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:22.594 [2024-12-07 05:40:55.699738] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.594 [2024-12-07 05:40:55.775304] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:22.594 [2024-12-07 05:40:55.775351] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:22.594 [2024-12-07 05:40:55.775367] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:13:22.594 [2024-12-07 05:40:55.775375] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:22.594 [2024-12-07 05:40:55.777557] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:22.594 [2024-12-07 05:40:55.777595] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:22.594 [2024-12-07 05:40:55.777678] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:22.594 [2024-12-07 05:40:55.777710] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:22.594 pt2 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:22.594 "name": "raid_bdev1", 00:13:22.594 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:22.594 "strip_size_kb": 64, 00:13:22.594 "state": "configuring", 00:13:22.594 "raid_level": "raid5f", 00:13:22.594 "superblock": true, 00:13:22.594 "num_base_bdevs": 3, 00:13:22.594 "num_base_bdevs_discovered": 1, 00:13:22.594 "num_base_bdevs_operational": 2, 00:13:22.594 "base_bdevs_list": [ 00:13:22.594 { 00:13:22.594 "name": null, 00:13:22.594 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:22.594 "is_configured": false, 00:13:22.594 "data_offset": 2048, 00:13:22.594 "data_size": 63488 00:13:22.594 }, 00:13:22.594 { 00:13:22.594 "name": "pt2", 00:13:22.594 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:22.594 "is_configured": true, 00:13:22.594 "data_offset": 2048, 00:13:22.594 "data_size": 63488 00:13:22.594 }, 00:13:22.594 { 00:13:22.594 "name": null, 00:13:22.594 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:22.594 "is_configured": false, 00:13:22.594 "data_offset": 2048, 00:13:22.594 "data_size": 63488 00:13:22.594 } 00:13:22.594 ] 00:13:22.594 }' 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:22.594 05:40:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.854 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:13:22.854 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:22.854 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:13:22.854 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:22.854 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.854 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.854 [2024-12-07 05:40:56.206586] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:22.854 [2024-12-07 05:40:56.206700] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:22.854 [2024-12-07 05:40:56.206746] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:13:22.854 [2024-12-07 05:40:56.206779] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:22.854 [2024-12-07 05:40:56.207165] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:22.854 [2024-12-07 05:40:56.207220] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:22.854 [2024-12-07 05:40:56.207319] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:22.854 [2024-12-07 05:40:56.207367] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:22.854 [2024-12-07 05:40:56.207478] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:22.854 [2024-12-07 05:40:56.207515] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:22.854 [2024-12-07 05:40:56.207780] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:13:22.854 [2024-12-07 05:40:56.208274] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:22.854 [2024-12-07 05:40:56.208331] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:13:22.854 [2024-12-07 05:40:56.208583] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:22.854 pt3 00:13:22.854 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.855 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.115 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.115 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.115 "name": "raid_bdev1", 00:13:23.115 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:23.115 "strip_size_kb": 64, 00:13:23.115 "state": "online", 00:13:23.115 "raid_level": "raid5f", 00:13:23.115 "superblock": true, 00:13:23.115 "num_base_bdevs": 3, 00:13:23.115 "num_base_bdevs_discovered": 2, 00:13:23.115 "num_base_bdevs_operational": 2, 00:13:23.115 "base_bdevs_list": [ 00:13:23.115 { 00:13:23.115 "name": null, 00:13:23.115 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.115 "is_configured": false, 00:13:23.115 "data_offset": 2048, 00:13:23.115 "data_size": 63488 00:13:23.115 }, 00:13:23.115 { 00:13:23.115 "name": "pt2", 00:13:23.115 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:23.115 "is_configured": true, 00:13:23.115 "data_offset": 2048, 00:13:23.115 "data_size": 63488 00:13:23.115 }, 00:13:23.115 { 00:13:23.115 "name": "pt3", 00:13:23.115 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:23.115 "is_configured": true, 00:13:23.115 "data_offset": 2048, 00:13:23.115 "data_size": 63488 00:13:23.115 } 00:13:23.115 ] 00:13:23.115 }' 00:13:23.115 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.115 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.375 [2024-12-07 05:40:56.609888] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:23.375 [2024-12-07 05:40:56.609912] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:23.375 [2024-12-07 05:40:56.609972] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:23.375 [2024-12-07 05:40:56.610025] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:23.375 [2024-12-07 05:40:56.610036] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.375 [2024-12-07 05:40:56.665790] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:23.375 [2024-12-07 05:40:56.665881] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:23.375 [2024-12-07 05:40:56.665913] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:13:23.375 [2024-12-07 05:40:56.665943] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:23.375 [2024-12-07 05:40:56.668133] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:23.375 [2024-12-07 05:40:56.668174] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:23.375 [2024-12-07 05:40:56.668234] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:23.375 [2024-12-07 05:40:56.668267] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:23.375 [2024-12-07 05:40:56.668368] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:13:23.375 [2024-12-07 05:40:56.668384] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:23.375 [2024-12-07 05:40:56.668399] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:13:23.375 [2024-12-07 05:40:56.668433] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:23.375 pt1 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.375 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.376 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.376 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.376 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.376 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.376 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.376 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.376 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.376 "name": "raid_bdev1", 00:13:23.376 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:23.376 "strip_size_kb": 64, 00:13:23.376 "state": "configuring", 00:13:23.376 "raid_level": "raid5f", 00:13:23.376 "superblock": true, 00:13:23.376 "num_base_bdevs": 3, 00:13:23.376 "num_base_bdevs_discovered": 1, 00:13:23.376 "num_base_bdevs_operational": 2, 00:13:23.376 "base_bdevs_list": [ 00:13:23.376 { 00:13:23.376 "name": null, 00:13:23.376 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.376 "is_configured": false, 00:13:23.376 "data_offset": 2048, 00:13:23.376 "data_size": 63488 00:13:23.376 }, 00:13:23.376 { 00:13:23.376 "name": "pt2", 00:13:23.376 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:23.376 "is_configured": true, 00:13:23.376 "data_offset": 2048, 00:13:23.376 "data_size": 63488 00:13:23.376 }, 00:13:23.376 { 00:13:23.376 "name": null, 00:13:23.376 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:23.376 "is_configured": false, 00:13:23.376 "data_offset": 2048, 00:13:23.376 "data_size": 63488 00:13:23.376 } 00:13:23.376 ] 00:13:23.376 }' 00:13:23.376 05:40:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.376 05:40:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.945 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:13:23.945 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.945 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.945 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:23.945 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.945 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:13:23.945 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.946 [2024-12-07 05:40:57.105052] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:23.946 [2024-12-07 05:40:57.105154] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:23.946 [2024-12-07 05:40:57.105188] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:13:23.946 [2024-12-07 05:40:57.105219] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:23.946 [2024-12-07 05:40:57.105602] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:23.946 [2024-12-07 05:40:57.105707] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:23.946 [2024-12-07 05:40:57.105815] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:23.946 [2024-12-07 05:40:57.105879] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:23.946 [2024-12-07 05:40:57.106012] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:13:23.946 [2024-12-07 05:40:57.106076] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:23.946 [2024-12-07 05:40:57.106353] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:13:23.946 [2024-12-07 05:40:57.106917] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:13:23.946 [2024-12-07 05:40:57.106977] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:13:23.946 [2024-12-07 05:40:57.107200] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:23.946 pt3 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.946 "name": "raid_bdev1", 00:13:23.946 "uuid": "681a9622-6617-42d6-8318-020dbde3bdd8", 00:13:23.946 "strip_size_kb": 64, 00:13:23.946 "state": "online", 00:13:23.946 "raid_level": "raid5f", 00:13:23.946 "superblock": true, 00:13:23.946 "num_base_bdevs": 3, 00:13:23.946 "num_base_bdevs_discovered": 2, 00:13:23.946 "num_base_bdevs_operational": 2, 00:13:23.946 "base_bdevs_list": [ 00:13:23.946 { 00:13:23.946 "name": null, 00:13:23.946 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.946 "is_configured": false, 00:13:23.946 "data_offset": 2048, 00:13:23.946 "data_size": 63488 00:13:23.946 }, 00:13:23.946 { 00:13:23.946 "name": "pt2", 00:13:23.946 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:23.946 "is_configured": true, 00:13:23.946 "data_offset": 2048, 00:13:23.946 "data_size": 63488 00:13:23.946 }, 00:13:23.946 { 00:13:23.946 "name": "pt3", 00:13:23.946 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:23.946 "is_configured": true, 00:13:23.946 "data_offset": 2048, 00:13:23.946 "data_size": 63488 00:13:23.946 } 00:13:23.946 ] 00:13:23.946 }' 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.946 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.206 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:24.206 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:13:24.206 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.206 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.206 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.206 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:13:24.206 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:13:24.206 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:24.206 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.206 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.465 [2024-12-07 05:40:57.576503] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 681a9622-6617-42d6-8318-020dbde3bdd8 '!=' 681a9622-6617-42d6-8318-020dbde3bdd8 ']' 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 91268 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 91268 ']' 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 91268 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 91268 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:24.465 killing process with pid 91268 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 91268' 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 91268 00:13:24.465 [2024-12-07 05:40:57.628157] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:24.465 [2024-12-07 05:40:57.628239] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:24.465 [2024-12-07 05:40:57.628299] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:24.465 [2024-12-07 05:40:57.628309] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:13:24.465 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 91268 00:13:24.465 [2024-12-07 05:40:57.661075] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:24.725 05:40:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:13:24.725 00:13:24.725 real 0m6.129s 00:13:24.725 user 0m10.293s 00:13:24.725 sys 0m1.276s 00:13:24.725 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:24.725 ************************************ 00:13:24.725 END TEST raid5f_superblock_test 00:13:24.725 ************************************ 00:13:24.725 05:40:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.725 05:40:57 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:13:24.725 05:40:57 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:13:24.725 05:40:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:24.725 05:40:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:24.725 05:40:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:24.725 ************************************ 00:13:24.725 START TEST raid5f_rebuild_test 00:13:24.725 ************************************ 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 false false true 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:24.725 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=91690 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 91690 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 91690 ']' 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:24.726 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:24.726 05:40:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.726 [2024-12-07 05:40:58.035385] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:13:24.726 [2024-12-07 05:40:58.035592] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:13:24.726 Zero copy mechanism will not be used. 00:13:24.726 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91690 ] 00:13:24.986 [2024-12-07 05:40:58.192310] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:24.986 [2024-12-07 05:40:58.218064] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:24.986 [2024-12-07 05:40:58.260847] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:24.986 [2024-12-07 05:40:58.260952] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.556 BaseBdev1_malloc 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.556 [2024-12-07 05:40:58.860164] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:25.556 [2024-12-07 05:40:58.860272] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:25.556 [2024-12-07 05:40:58.860306] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:25.556 [2024-12-07 05:40:58.860318] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:25.556 [2024-12-07 05:40:58.862406] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:25.556 [2024-12-07 05:40:58.862447] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:25.556 BaseBdev1 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.556 BaseBdev2_malloc 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.556 [2024-12-07 05:40:58.880576] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:25.556 [2024-12-07 05:40:58.880681] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:25.556 [2024-12-07 05:40:58.880724] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:25.556 [2024-12-07 05:40:58.880752] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:25.556 [2024-12-07 05:40:58.882805] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:25.556 [2024-12-07 05:40:58.882901] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:25.556 BaseBdev2 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.556 BaseBdev3_malloc 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.556 [2024-12-07 05:40:58.908950] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:25.556 [2024-12-07 05:40:58.909000] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:25.556 [2024-12-07 05:40:58.909023] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:25.556 [2024-12-07 05:40:58.909031] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:25.556 [2024-12-07 05:40:58.911135] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:25.556 [2024-12-07 05:40:58.911172] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:25.556 BaseBdev3 00:13:25.556 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.557 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:25.557 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.557 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.817 spare_malloc 00:13:25.817 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.817 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:25.817 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.817 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.817 spare_delay 00:13:25.817 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.817 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:25.817 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.817 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.817 [2024-12-07 05:40:58.960411] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:25.817 [2024-12-07 05:40:58.960461] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:25.817 [2024-12-07 05:40:58.960487] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:13:25.817 [2024-12-07 05:40:58.960496] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:25.817 [2024-12-07 05:40:58.962595] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:25.817 [2024-12-07 05:40:58.962638] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:25.817 spare 00:13:25.817 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.817 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.818 [2024-12-07 05:40:58.972460] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:25.818 [2024-12-07 05:40:58.974249] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:25.818 [2024-12-07 05:40:58.974359] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:25.818 [2024-12-07 05:40:58.974447] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:25.818 [2024-12-07 05:40:58.974459] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:25.818 [2024-12-07 05:40:58.974731] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:25.818 [2024-12-07 05:40:58.975129] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:25.818 [2024-12-07 05:40:58.975140] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:25.818 [2024-12-07 05:40:58.975261] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.818 05:40:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.818 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.818 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:25.818 "name": "raid_bdev1", 00:13:25.818 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:25.818 "strip_size_kb": 64, 00:13:25.818 "state": "online", 00:13:25.818 "raid_level": "raid5f", 00:13:25.818 "superblock": false, 00:13:25.818 "num_base_bdevs": 3, 00:13:25.818 "num_base_bdevs_discovered": 3, 00:13:25.818 "num_base_bdevs_operational": 3, 00:13:25.818 "base_bdevs_list": [ 00:13:25.818 { 00:13:25.818 "name": "BaseBdev1", 00:13:25.818 "uuid": "0ca55bba-f951-51a2-9150-6fd22faa5830", 00:13:25.818 "is_configured": true, 00:13:25.818 "data_offset": 0, 00:13:25.818 "data_size": 65536 00:13:25.818 }, 00:13:25.818 { 00:13:25.818 "name": "BaseBdev2", 00:13:25.818 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:25.818 "is_configured": true, 00:13:25.818 "data_offset": 0, 00:13:25.818 "data_size": 65536 00:13:25.818 }, 00:13:25.818 { 00:13:25.818 "name": "BaseBdev3", 00:13:25.818 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:25.818 "is_configured": true, 00:13:25.818 "data_offset": 0, 00:13:25.818 "data_size": 65536 00:13:25.818 } 00:13:25.818 ] 00:13:25.818 }' 00:13:25.818 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:25.818 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.077 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:26.077 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:26.077 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.077 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.337 [2024-12-07 05:40:59.447989] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:26.337 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:26.337 [2024-12-07 05:40:59.695552] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:13:26.614 /dev/nbd0 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:26.614 1+0 records in 00:13:26.614 1+0 records out 00:13:26.614 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000354552 s, 11.6 MB/s 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:13:26.614 05:40:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:13:26.873 512+0 records in 00:13:26.873 512+0 records out 00:13:26.873 67108864 bytes (67 MB, 64 MiB) copied, 0.273575 s, 245 MB/s 00:13:26.873 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:26.873 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:26.873 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:26.873 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:26.873 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:26.873 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:26.873 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:26.873 [2024-12-07 05:41:00.226152] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.133 [2024-12-07 05:41:00.262155] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:27.133 "name": "raid_bdev1", 00:13:27.133 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:27.133 "strip_size_kb": 64, 00:13:27.133 "state": "online", 00:13:27.133 "raid_level": "raid5f", 00:13:27.133 "superblock": false, 00:13:27.133 "num_base_bdevs": 3, 00:13:27.133 "num_base_bdevs_discovered": 2, 00:13:27.133 "num_base_bdevs_operational": 2, 00:13:27.133 "base_bdevs_list": [ 00:13:27.133 { 00:13:27.133 "name": null, 00:13:27.133 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:27.133 "is_configured": false, 00:13:27.133 "data_offset": 0, 00:13:27.133 "data_size": 65536 00:13:27.133 }, 00:13:27.133 { 00:13:27.133 "name": "BaseBdev2", 00:13:27.133 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:27.133 "is_configured": true, 00:13:27.133 "data_offset": 0, 00:13:27.133 "data_size": 65536 00:13:27.133 }, 00:13:27.133 { 00:13:27.133 "name": "BaseBdev3", 00:13:27.133 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:27.133 "is_configured": true, 00:13:27.133 "data_offset": 0, 00:13:27.133 "data_size": 65536 00:13:27.133 } 00:13:27.133 ] 00:13:27.133 }' 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:27.133 05:41:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.393 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:27.393 05:41:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.393 05:41:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.393 [2024-12-07 05:41:00.685476] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:27.393 [2024-12-07 05:41:00.690031] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027cd0 00:13:27.393 05:41:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.393 05:41:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:27.393 [2024-12-07 05:41:00.692231] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:28.772 "name": "raid_bdev1", 00:13:28.772 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:28.772 "strip_size_kb": 64, 00:13:28.772 "state": "online", 00:13:28.772 "raid_level": "raid5f", 00:13:28.772 "superblock": false, 00:13:28.772 "num_base_bdevs": 3, 00:13:28.772 "num_base_bdevs_discovered": 3, 00:13:28.772 "num_base_bdevs_operational": 3, 00:13:28.772 "process": { 00:13:28.772 "type": "rebuild", 00:13:28.772 "target": "spare", 00:13:28.772 "progress": { 00:13:28.772 "blocks": 20480, 00:13:28.772 "percent": 15 00:13:28.772 } 00:13:28.772 }, 00:13:28.772 "base_bdevs_list": [ 00:13:28.772 { 00:13:28.772 "name": "spare", 00:13:28.772 "uuid": "c524bab9-0294-57ff-85c0-572ef4702a05", 00:13:28.772 "is_configured": true, 00:13:28.772 "data_offset": 0, 00:13:28.772 "data_size": 65536 00:13:28.772 }, 00:13:28.772 { 00:13:28.772 "name": "BaseBdev2", 00:13:28.772 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:28.772 "is_configured": true, 00:13:28.772 "data_offset": 0, 00:13:28.772 "data_size": 65536 00:13:28.772 }, 00:13:28.772 { 00:13:28.772 "name": "BaseBdev3", 00:13:28.772 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:28.772 "is_configured": true, 00:13:28.772 "data_offset": 0, 00:13:28.772 "data_size": 65536 00:13:28.772 } 00:13:28.772 ] 00:13:28.772 }' 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.772 [2024-12-07 05:41:01.832595] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:28.772 [2024-12-07 05:41:01.899455] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:28.772 [2024-12-07 05:41:01.899514] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:28.772 [2024-12-07 05:41:01.899530] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:28.772 [2024-12-07 05:41:01.899552] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:28.772 "name": "raid_bdev1", 00:13:28.772 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:28.772 "strip_size_kb": 64, 00:13:28.772 "state": "online", 00:13:28.772 "raid_level": "raid5f", 00:13:28.772 "superblock": false, 00:13:28.772 "num_base_bdevs": 3, 00:13:28.772 "num_base_bdevs_discovered": 2, 00:13:28.772 "num_base_bdevs_operational": 2, 00:13:28.772 "base_bdevs_list": [ 00:13:28.772 { 00:13:28.772 "name": null, 00:13:28.772 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.772 "is_configured": false, 00:13:28.772 "data_offset": 0, 00:13:28.772 "data_size": 65536 00:13:28.772 }, 00:13:28.772 { 00:13:28.772 "name": "BaseBdev2", 00:13:28.772 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:28.772 "is_configured": true, 00:13:28.772 "data_offset": 0, 00:13:28.772 "data_size": 65536 00:13:28.772 }, 00:13:28.772 { 00:13:28.772 "name": "BaseBdev3", 00:13:28.772 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:28.772 "is_configured": true, 00:13:28.772 "data_offset": 0, 00:13:28.772 "data_size": 65536 00:13:28.772 } 00:13:28.772 ] 00:13:28.772 }' 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:28.772 05:41:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.032 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:29.032 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:29.032 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:29.033 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:29.033 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:29.033 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.033 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:29.033 05:41:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.033 05:41:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.033 05:41:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.033 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:29.033 "name": "raid_bdev1", 00:13:29.033 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:29.033 "strip_size_kb": 64, 00:13:29.033 "state": "online", 00:13:29.033 "raid_level": "raid5f", 00:13:29.033 "superblock": false, 00:13:29.033 "num_base_bdevs": 3, 00:13:29.033 "num_base_bdevs_discovered": 2, 00:13:29.033 "num_base_bdevs_operational": 2, 00:13:29.033 "base_bdevs_list": [ 00:13:29.033 { 00:13:29.033 "name": null, 00:13:29.033 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:29.033 "is_configured": false, 00:13:29.033 "data_offset": 0, 00:13:29.033 "data_size": 65536 00:13:29.033 }, 00:13:29.033 { 00:13:29.033 "name": "BaseBdev2", 00:13:29.033 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:29.033 "is_configured": true, 00:13:29.033 "data_offset": 0, 00:13:29.033 "data_size": 65536 00:13:29.033 }, 00:13:29.033 { 00:13:29.033 "name": "BaseBdev3", 00:13:29.033 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:29.033 "is_configured": true, 00:13:29.033 "data_offset": 0, 00:13:29.033 "data_size": 65536 00:13:29.033 } 00:13:29.033 ] 00:13:29.033 }' 00:13:29.033 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:29.293 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:29.293 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:29.293 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:29.293 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:29.293 05:41:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.293 05:41:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.293 [2024-12-07 05:41:02.500413] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:29.293 [2024-12-07 05:41:02.504615] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:13:29.293 05:41:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.293 05:41:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:29.293 [2024-12-07 05:41:02.506767] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:30.233 "name": "raid_bdev1", 00:13:30.233 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:30.233 "strip_size_kb": 64, 00:13:30.233 "state": "online", 00:13:30.233 "raid_level": "raid5f", 00:13:30.233 "superblock": false, 00:13:30.233 "num_base_bdevs": 3, 00:13:30.233 "num_base_bdevs_discovered": 3, 00:13:30.233 "num_base_bdevs_operational": 3, 00:13:30.233 "process": { 00:13:30.233 "type": "rebuild", 00:13:30.233 "target": "spare", 00:13:30.233 "progress": { 00:13:30.233 "blocks": 20480, 00:13:30.233 "percent": 15 00:13:30.233 } 00:13:30.233 }, 00:13:30.233 "base_bdevs_list": [ 00:13:30.233 { 00:13:30.233 "name": "spare", 00:13:30.233 "uuid": "c524bab9-0294-57ff-85c0-572ef4702a05", 00:13:30.233 "is_configured": true, 00:13:30.233 "data_offset": 0, 00:13:30.233 "data_size": 65536 00:13:30.233 }, 00:13:30.233 { 00:13:30.233 "name": "BaseBdev2", 00:13:30.233 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:30.233 "is_configured": true, 00:13:30.233 "data_offset": 0, 00:13:30.233 "data_size": 65536 00:13:30.233 }, 00:13:30.233 { 00:13:30.233 "name": "BaseBdev3", 00:13:30.233 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:30.233 "is_configured": true, 00:13:30.233 "data_offset": 0, 00:13:30.233 "data_size": 65536 00:13:30.233 } 00:13:30.233 ] 00:13:30.233 }' 00:13:30.233 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=440 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.493 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:30.493 "name": "raid_bdev1", 00:13:30.493 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:30.493 "strip_size_kb": 64, 00:13:30.493 "state": "online", 00:13:30.493 "raid_level": "raid5f", 00:13:30.493 "superblock": false, 00:13:30.493 "num_base_bdevs": 3, 00:13:30.493 "num_base_bdevs_discovered": 3, 00:13:30.493 "num_base_bdevs_operational": 3, 00:13:30.493 "process": { 00:13:30.493 "type": "rebuild", 00:13:30.493 "target": "spare", 00:13:30.493 "progress": { 00:13:30.493 "blocks": 22528, 00:13:30.493 "percent": 17 00:13:30.493 } 00:13:30.493 }, 00:13:30.493 "base_bdevs_list": [ 00:13:30.493 { 00:13:30.494 "name": "spare", 00:13:30.494 "uuid": "c524bab9-0294-57ff-85c0-572ef4702a05", 00:13:30.494 "is_configured": true, 00:13:30.494 "data_offset": 0, 00:13:30.494 "data_size": 65536 00:13:30.494 }, 00:13:30.494 { 00:13:30.494 "name": "BaseBdev2", 00:13:30.494 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:30.494 "is_configured": true, 00:13:30.494 "data_offset": 0, 00:13:30.494 "data_size": 65536 00:13:30.494 }, 00:13:30.494 { 00:13:30.494 "name": "BaseBdev3", 00:13:30.494 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:30.494 "is_configured": true, 00:13:30.494 "data_offset": 0, 00:13:30.494 "data_size": 65536 00:13:30.494 } 00:13:30.494 ] 00:13:30.494 }' 00:13:30.494 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:30.494 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:30.494 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:30.494 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:30.494 05:41:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:31.435 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:31.435 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:31.435 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:31.435 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:31.435 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:31.435 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:31.696 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:31.696 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:31.696 05:41:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:31.696 05:41:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.696 05:41:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:31.696 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:31.696 "name": "raid_bdev1", 00:13:31.696 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:31.696 "strip_size_kb": 64, 00:13:31.696 "state": "online", 00:13:31.696 "raid_level": "raid5f", 00:13:31.696 "superblock": false, 00:13:31.696 "num_base_bdevs": 3, 00:13:31.696 "num_base_bdevs_discovered": 3, 00:13:31.696 "num_base_bdevs_operational": 3, 00:13:31.696 "process": { 00:13:31.696 "type": "rebuild", 00:13:31.696 "target": "spare", 00:13:31.696 "progress": { 00:13:31.696 "blocks": 45056, 00:13:31.696 "percent": 34 00:13:31.696 } 00:13:31.696 }, 00:13:31.696 "base_bdevs_list": [ 00:13:31.696 { 00:13:31.696 "name": "spare", 00:13:31.696 "uuid": "c524bab9-0294-57ff-85c0-572ef4702a05", 00:13:31.696 "is_configured": true, 00:13:31.696 "data_offset": 0, 00:13:31.696 "data_size": 65536 00:13:31.696 }, 00:13:31.696 { 00:13:31.696 "name": "BaseBdev2", 00:13:31.696 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:31.696 "is_configured": true, 00:13:31.696 "data_offset": 0, 00:13:31.696 "data_size": 65536 00:13:31.696 }, 00:13:31.696 { 00:13:31.696 "name": "BaseBdev3", 00:13:31.696 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:31.696 "is_configured": true, 00:13:31.696 "data_offset": 0, 00:13:31.696 "data_size": 65536 00:13:31.696 } 00:13:31.696 ] 00:13:31.696 }' 00:13:31.696 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:31.696 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:31.696 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:31.696 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:31.696 05:41:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:32.632 05:41:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:32.632 05:41:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:32.632 05:41:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:32.632 05:41:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:32.632 05:41:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:32.632 05:41:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:32.632 05:41:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.632 05:41:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:32.632 05:41:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.632 05:41:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.632 05:41:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.892 05:41:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:32.892 "name": "raid_bdev1", 00:13:32.892 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:32.892 "strip_size_kb": 64, 00:13:32.892 "state": "online", 00:13:32.892 "raid_level": "raid5f", 00:13:32.892 "superblock": false, 00:13:32.892 "num_base_bdevs": 3, 00:13:32.892 "num_base_bdevs_discovered": 3, 00:13:32.892 "num_base_bdevs_operational": 3, 00:13:32.892 "process": { 00:13:32.892 "type": "rebuild", 00:13:32.892 "target": "spare", 00:13:32.892 "progress": { 00:13:32.892 "blocks": 69632, 00:13:32.892 "percent": 53 00:13:32.892 } 00:13:32.892 }, 00:13:32.892 "base_bdevs_list": [ 00:13:32.892 { 00:13:32.892 "name": "spare", 00:13:32.892 "uuid": "c524bab9-0294-57ff-85c0-572ef4702a05", 00:13:32.892 "is_configured": true, 00:13:32.892 "data_offset": 0, 00:13:32.892 "data_size": 65536 00:13:32.892 }, 00:13:32.892 { 00:13:32.892 "name": "BaseBdev2", 00:13:32.892 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:32.892 "is_configured": true, 00:13:32.892 "data_offset": 0, 00:13:32.892 "data_size": 65536 00:13:32.892 }, 00:13:32.892 { 00:13:32.892 "name": "BaseBdev3", 00:13:32.892 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:32.892 "is_configured": true, 00:13:32.892 "data_offset": 0, 00:13:32.892 "data_size": 65536 00:13:32.892 } 00:13:32.892 ] 00:13:32.892 }' 00:13:32.892 05:41:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:32.892 05:41:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:32.892 05:41:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:32.892 05:41:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:32.892 05:41:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:33.851 "name": "raid_bdev1", 00:13:33.851 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:33.851 "strip_size_kb": 64, 00:13:33.851 "state": "online", 00:13:33.851 "raid_level": "raid5f", 00:13:33.851 "superblock": false, 00:13:33.851 "num_base_bdevs": 3, 00:13:33.851 "num_base_bdevs_discovered": 3, 00:13:33.851 "num_base_bdevs_operational": 3, 00:13:33.851 "process": { 00:13:33.851 "type": "rebuild", 00:13:33.851 "target": "spare", 00:13:33.851 "progress": { 00:13:33.851 "blocks": 92160, 00:13:33.851 "percent": 70 00:13:33.851 } 00:13:33.851 }, 00:13:33.851 "base_bdevs_list": [ 00:13:33.851 { 00:13:33.851 "name": "spare", 00:13:33.851 "uuid": "c524bab9-0294-57ff-85c0-572ef4702a05", 00:13:33.851 "is_configured": true, 00:13:33.851 "data_offset": 0, 00:13:33.851 "data_size": 65536 00:13:33.851 }, 00:13:33.851 { 00:13:33.851 "name": "BaseBdev2", 00:13:33.851 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:33.851 "is_configured": true, 00:13:33.851 "data_offset": 0, 00:13:33.851 "data_size": 65536 00:13:33.851 }, 00:13:33.851 { 00:13:33.851 "name": "BaseBdev3", 00:13:33.851 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:33.851 "is_configured": true, 00:13:33.851 "data_offset": 0, 00:13:33.851 "data_size": 65536 00:13:33.851 } 00:13:33.851 ] 00:13:33.851 }' 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:33.851 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:34.111 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:34.111 05:41:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:35.052 "name": "raid_bdev1", 00:13:35.052 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:35.052 "strip_size_kb": 64, 00:13:35.052 "state": "online", 00:13:35.052 "raid_level": "raid5f", 00:13:35.052 "superblock": false, 00:13:35.052 "num_base_bdevs": 3, 00:13:35.052 "num_base_bdevs_discovered": 3, 00:13:35.052 "num_base_bdevs_operational": 3, 00:13:35.052 "process": { 00:13:35.052 "type": "rebuild", 00:13:35.052 "target": "spare", 00:13:35.052 "progress": { 00:13:35.052 "blocks": 116736, 00:13:35.052 "percent": 89 00:13:35.052 } 00:13:35.052 }, 00:13:35.052 "base_bdevs_list": [ 00:13:35.052 { 00:13:35.052 "name": "spare", 00:13:35.052 "uuid": "c524bab9-0294-57ff-85c0-572ef4702a05", 00:13:35.052 "is_configured": true, 00:13:35.052 "data_offset": 0, 00:13:35.052 "data_size": 65536 00:13:35.052 }, 00:13:35.052 { 00:13:35.052 "name": "BaseBdev2", 00:13:35.052 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:35.052 "is_configured": true, 00:13:35.052 "data_offset": 0, 00:13:35.052 "data_size": 65536 00:13:35.052 }, 00:13:35.052 { 00:13:35.052 "name": "BaseBdev3", 00:13:35.052 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:35.052 "is_configured": true, 00:13:35.052 "data_offset": 0, 00:13:35.052 "data_size": 65536 00:13:35.052 } 00:13:35.052 ] 00:13:35.052 }' 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:35.052 05:41:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:35.622 [2024-12-07 05:41:08.942246] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:35.622 [2024-12-07 05:41:08.942310] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:35.622 [2024-12-07 05:41:08.942367] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.191 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:36.191 "name": "raid_bdev1", 00:13:36.191 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:36.191 "strip_size_kb": 64, 00:13:36.191 "state": "online", 00:13:36.191 "raid_level": "raid5f", 00:13:36.191 "superblock": false, 00:13:36.191 "num_base_bdevs": 3, 00:13:36.191 "num_base_bdevs_discovered": 3, 00:13:36.191 "num_base_bdevs_operational": 3, 00:13:36.191 "base_bdevs_list": [ 00:13:36.191 { 00:13:36.191 "name": "spare", 00:13:36.191 "uuid": "c524bab9-0294-57ff-85c0-572ef4702a05", 00:13:36.191 "is_configured": true, 00:13:36.191 "data_offset": 0, 00:13:36.191 "data_size": 65536 00:13:36.191 }, 00:13:36.191 { 00:13:36.191 "name": "BaseBdev2", 00:13:36.192 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:36.192 "is_configured": true, 00:13:36.192 "data_offset": 0, 00:13:36.192 "data_size": 65536 00:13:36.192 }, 00:13:36.192 { 00:13:36.192 "name": "BaseBdev3", 00:13:36.192 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:36.192 "is_configured": true, 00:13:36.192 "data_offset": 0, 00:13:36.192 "data_size": 65536 00:13:36.192 } 00:13:36.192 ] 00:13:36.192 }' 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.192 05:41:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:36.452 "name": "raid_bdev1", 00:13:36.452 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:36.452 "strip_size_kb": 64, 00:13:36.452 "state": "online", 00:13:36.452 "raid_level": "raid5f", 00:13:36.452 "superblock": false, 00:13:36.452 "num_base_bdevs": 3, 00:13:36.452 "num_base_bdevs_discovered": 3, 00:13:36.452 "num_base_bdevs_operational": 3, 00:13:36.452 "base_bdevs_list": [ 00:13:36.452 { 00:13:36.452 "name": "spare", 00:13:36.452 "uuid": "c524bab9-0294-57ff-85c0-572ef4702a05", 00:13:36.452 "is_configured": true, 00:13:36.452 "data_offset": 0, 00:13:36.452 "data_size": 65536 00:13:36.452 }, 00:13:36.452 { 00:13:36.452 "name": "BaseBdev2", 00:13:36.452 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:36.452 "is_configured": true, 00:13:36.452 "data_offset": 0, 00:13:36.452 "data_size": 65536 00:13:36.452 }, 00:13:36.452 { 00:13:36.452 "name": "BaseBdev3", 00:13:36.452 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:36.452 "is_configured": true, 00:13:36.452 "data_offset": 0, 00:13:36.452 "data_size": 65536 00:13:36.452 } 00:13:36.452 ] 00:13:36.452 }' 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:36.452 "name": "raid_bdev1", 00:13:36.452 "uuid": "bfcad438-2160-4425-a8a7-63a16dc69220", 00:13:36.452 "strip_size_kb": 64, 00:13:36.452 "state": "online", 00:13:36.452 "raid_level": "raid5f", 00:13:36.452 "superblock": false, 00:13:36.452 "num_base_bdevs": 3, 00:13:36.452 "num_base_bdevs_discovered": 3, 00:13:36.452 "num_base_bdevs_operational": 3, 00:13:36.452 "base_bdevs_list": [ 00:13:36.452 { 00:13:36.452 "name": "spare", 00:13:36.452 "uuid": "c524bab9-0294-57ff-85c0-572ef4702a05", 00:13:36.452 "is_configured": true, 00:13:36.452 "data_offset": 0, 00:13:36.452 "data_size": 65536 00:13:36.452 }, 00:13:36.452 { 00:13:36.452 "name": "BaseBdev2", 00:13:36.452 "uuid": "69eaa0fa-6407-512d-ab36-eea181355826", 00:13:36.452 "is_configured": true, 00:13:36.452 "data_offset": 0, 00:13:36.452 "data_size": 65536 00:13:36.452 }, 00:13:36.452 { 00:13:36.452 "name": "BaseBdev3", 00:13:36.452 "uuid": "76d4664c-75f0-5865-b94a-442e740b5695", 00:13:36.452 "is_configured": true, 00:13:36.452 "data_offset": 0, 00:13:36.452 "data_size": 65536 00:13:36.452 } 00:13:36.452 ] 00:13:36.452 }' 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:36.452 05:41:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.712 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:36.712 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.712 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.972 [2024-12-07 05:41:10.081827] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:36.972 [2024-12-07 05:41:10.081914] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:36.972 [2024-12-07 05:41:10.082020] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:36.972 [2024-12-07 05:41:10.082141] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:36.972 [2024-12-07 05:41:10.082154] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:36.972 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:37.232 /dev/nbd0 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:37.232 1+0 records in 00:13:37.232 1+0 records out 00:13:37.232 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000316889 s, 12.9 MB/s 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:37.232 /dev/nbd1 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:37.232 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:37.492 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:37.493 1+0 records in 00:13:37.493 1+0 records out 00:13:37.493 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.0003491 s, 11.7 MB/s 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:37.493 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:37.753 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:37.753 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:37.753 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:37.753 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:37.753 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:37.753 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:37.753 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:37.753 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:37.753 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:37.753 05:41:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 91690 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 91690 ']' 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 91690 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 91690 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 91690' 00:13:38.013 killing process with pid 91690 00:13:38.013 Received shutdown signal, test time was about 60.000000 seconds 00:13:38.013 00:13:38.013 Latency(us) 00:13:38.013 [2024-12-07T05:41:11.381Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:38.013 [2024-12-07T05:41:11.381Z] =================================================================================================================== 00:13:38.013 [2024-12-07T05:41:11.381Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 91690 00:13:38.013 [2024-12-07 05:41:11.170514] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:38.013 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 91690 00:13:38.013 [2024-12-07 05:41:11.210700] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:38.273 05:41:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:13:38.273 00:13:38.273 real 0m13.463s 00:13:38.274 user 0m16.943s 00:13:38.274 sys 0m1.791s 00:13:38.274 ************************************ 00:13:38.274 END TEST raid5f_rebuild_test 00:13:38.274 ************************************ 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.274 05:41:11 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:13:38.274 05:41:11 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:38.274 05:41:11 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:38.274 05:41:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:38.274 ************************************ 00:13:38.274 START TEST raid5f_rebuild_test_sb 00:13:38.274 ************************************ 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 true false true 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=92118 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 92118 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 92118 ']' 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:38.274 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:38.274 05:41:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.274 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:38.274 Zero copy mechanism will not be used. 00:13:38.274 [2024-12-07 05:41:11.569095] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:13:38.274 [2024-12-07 05:41:11.569213] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92118 ] 00:13:38.534 [2024-12-07 05:41:11.724313] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:38.534 [2024-12-07 05:41:11.748936] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:38.534 [2024-12-07 05:41:11.790894] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:38.534 [2024-12-07 05:41:11.790929] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.105 BaseBdev1_malloc 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.105 [2024-12-07 05:41:12.390051] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:39.105 [2024-12-07 05:41:12.390114] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:39.105 [2024-12-07 05:41:12.390143] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:39.105 [2024-12-07 05:41:12.390155] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:39.105 [2024-12-07 05:41:12.392216] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:39.105 [2024-12-07 05:41:12.392251] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:39.105 BaseBdev1 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.105 BaseBdev2_malloc 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.105 [2024-12-07 05:41:12.410583] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:39.105 [2024-12-07 05:41:12.410662] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:39.105 [2024-12-07 05:41:12.410686] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:39.105 [2024-12-07 05:41:12.410695] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:39.105 [2024-12-07 05:41:12.412762] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:39.105 [2024-12-07 05:41:12.412865] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:39.105 BaseBdev2 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.105 BaseBdev3_malloc 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.105 [2024-12-07 05:41:12.435242] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:39.105 [2024-12-07 05:41:12.435297] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:39.105 [2024-12-07 05:41:12.435320] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:39.105 [2024-12-07 05:41:12.435329] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:39.105 [2024-12-07 05:41:12.437445] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:39.105 [2024-12-07 05:41:12.437494] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:39.105 BaseBdev3 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.105 spare_malloc 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.105 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.366 spare_delay 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.366 [2024-12-07 05:41:12.485501] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:39.366 [2024-12-07 05:41:12.485555] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:39.366 [2024-12-07 05:41:12.485580] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:13:39.366 [2024-12-07 05:41:12.485589] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:39.366 [2024-12-07 05:41:12.487680] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:39.366 [2024-12-07 05:41:12.487767] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:39.366 spare 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.366 [2024-12-07 05:41:12.497553] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:39.366 [2024-12-07 05:41:12.499356] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:39.366 [2024-12-07 05:41:12.499415] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:39.366 [2024-12-07 05:41:12.499567] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:39.366 [2024-12-07 05:41:12.499581] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:39.366 [2024-12-07 05:41:12.499835] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:39.366 [2024-12-07 05:41:12.500219] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:39.366 [2024-12-07 05:41:12.500236] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:39.366 [2024-12-07 05:41:12.500352] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:39.366 "name": "raid_bdev1", 00:13:39.366 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:39.366 "strip_size_kb": 64, 00:13:39.366 "state": "online", 00:13:39.366 "raid_level": "raid5f", 00:13:39.366 "superblock": true, 00:13:39.366 "num_base_bdevs": 3, 00:13:39.366 "num_base_bdevs_discovered": 3, 00:13:39.366 "num_base_bdevs_operational": 3, 00:13:39.366 "base_bdevs_list": [ 00:13:39.366 { 00:13:39.366 "name": "BaseBdev1", 00:13:39.366 "uuid": "56d7e18e-7e88-5556-ac6e-2eff25cb2a23", 00:13:39.366 "is_configured": true, 00:13:39.366 "data_offset": 2048, 00:13:39.366 "data_size": 63488 00:13:39.366 }, 00:13:39.366 { 00:13:39.366 "name": "BaseBdev2", 00:13:39.366 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:39.366 "is_configured": true, 00:13:39.366 "data_offset": 2048, 00:13:39.366 "data_size": 63488 00:13:39.366 }, 00:13:39.366 { 00:13:39.366 "name": "BaseBdev3", 00:13:39.366 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:39.366 "is_configured": true, 00:13:39.366 "data_offset": 2048, 00:13:39.366 "data_size": 63488 00:13:39.366 } 00:13:39.366 ] 00:13:39.366 }' 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:39.366 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.626 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:39.626 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:39.626 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.626 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.626 [2024-12-07 05:41:12.933116] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:39.626 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.627 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:13:39.627 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:39.627 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:39.627 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.627 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.887 05:41:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:39.887 [2024-12-07 05:41:13.216715] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:13:39.887 /dev/nbd0 00:13:39.887 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:40.147 1+0 records in 00:13:40.147 1+0 records out 00:13:40.147 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000210933 s, 19.4 MB/s 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:13:40.147 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:13:40.407 496+0 records in 00:13:40.407 496+0 records out 00:13:40.407 65011712 bytes (65 MB, 62 MiB) copied, 0.29214 s, 223 MB/s 00:13:40.407 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:40.407 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:40.407 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:40.407 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:40.407 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:40.407 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:40.407 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:40.667 [2024-12-07 05:41:13.790709] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:40.667 [2024-12-07 05:41:13.810753] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:40.667 "name": "raid_bdev1", 00:13:40.667 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:40.667 "strip_size_kb": 64, 00:13:40.667 "state": "online", 00:13:40.667 "raid_level": "raid5f", 00:13:40.667 "superblock": true, 00:13:40.667 "num_base_bdevs": 3, 00:13:40.667 "num_base_bdevs_discovered": 2, 00:13:40.667 "num_base_bdevs_operational": 2, 00:13:40.667 "base_bdevs_list": [ 00:13:40.667 { 00:13:40.667 "name": null, 00:13:40.667 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:40.667 "is_configured": false, 00:13:40.667 "data_offset": 0, 00:13:40.667 "data_size": 63488 00:13:40.667 }, 00:13:40.667 { 00:13:40.667 "name": "BaseBdev2", 00:13:40.667 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:40.667 "is_configured": true, 00:13:40.667 "data_offset": 2048, 00:13:40.667 "data_size": 63488 00:13:40.667 }, 00:13:40.667 { 00:13:40.667 "name": "BaseBdev3", 00:13:40.667 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:40.667 "is_configured": true, 00:13:40.667 "data_offset": 2048, 00:13:40.667 "data_size": 63488 00:13:40.667 } 00:13:40.667 ] 00:13:40.667 }' 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:40.667 05:41:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:40.927 05:41:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:40.927 05:41:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.927 05:41:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:40.927 [2024-12-07 05:41:14.226076] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:40.927 [2024-12-07 05:41:14.230668] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000255d0 00:13:40.927 05:41:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.927 05:41:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:40.927 [2024-12-07 05:41:14.232886] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:42.308 "name": "raid_bdev1", 00:13:42.308 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:42.308 "strip_size_kb": 64, 00:13:42.308 "state": "online", 00:13:42.308 "raid_level": "raid5f", 00:13:42.308 "superblock": true, 00:13:42.308 "num_base_bdevs": 3, 00:13:42.308 "num_base_bdevs_discovered": 3, 00:13:42.308 "num_base_bdevs_operational": 3, 00:13:42.308 "process": { 00:13:42.308 "type": "rebuild", 00:13:42.308 "target": "spare", 00:13:42.308 "progress": { 00:13:42.308 "blocks": 20480, 00:13:42.308 "percent": 16 00:13:42.308 } 00:13:42.308 }, 00:13:42.308 "base_bdevs_list": [ 00:13:42.308 { 00:13:42.308 "name": "spare", 00:13:42.308 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:42.308 "is_configured": true, 00:13:42.308 "data_offset": 2048, 00:13:42.308 "data_size": 63488 00:13:42.308 }, 00:13:42.308 { 00:13:42.308 "name": "BaseBdev2", 00:13:42.308 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:42.308 "is_configured": true, 00:13:42.308 "data_offset": 2048, 00:13:42.308 "data_size": 63488 00:13:42.308 }, 00:13:42.308 { 00:13:42.308 "name": "BaseBdev3", 00:13:42.308 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:42.308 "is_configured": true, 00:13:42.308 "data_offset": 2048, 00:13:42.308 "data_size": 63488 00:13:42.308 } 00:13:42.308 ] 00:13:42.308 }' 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.308 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.309 [2024-12-07 05:41:15.396433] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:42.309 [2024-12-07 05:41:15.440171] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:42.309 [2024-12-07 05:41:15.440250] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:42.309 [2024-12-07 05:41:15.440266] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:42.309 [2024-12-07 05:41:15.440276] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:42.309 "name": "raid_bdev1", 00:13:42.309 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:42.309 "strip_size_kb": 64, 00:13:42.309 "state": "online", 00:13:42.309 "raid_level": "raid5f", 00:13:42.309 "superblock": true, 00:13:42.309 "num_base_bdevs": 3, 00:13:42.309 "num_base_bdevs_discovered": 2, 00:13:42.309 "num_base_bdevs_operational": 2, 00:13:42.309 "base_bdevs_list": [ 00:13:42.309 { 00:13:42.309 "name": null, 00:13:42.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.309 "is_configured": false, 00:13:42.309 "data_offset": 0, 00:13:42.309 "data_size": 63488 00:13:42.309 }, 00:13:42.309 { 00:13:42.309 "name": "BaseBdev2", 00:13:42.309 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:42.309 "is_configured": true, 00:13:42.309 "data_offset": 2048, 00:13:42.309 "data_size": 63488 00:13:42.309 }, 00:13:42.309 { 00:13:42.309 "name": "BaseBdev3", 00:13:42.309 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:42.309 "is_configured": true, 00:13:42.309 "data_offset": 2048, 00:13:42.309 "data_size": 63488 00:13:42.309 } 00:13:42.309 ] 00:13:42.309 }' 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:42.309 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.568 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:42.568 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:42.568 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:42.568 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:42.568 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:42.568 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.568 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:42.568 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.568 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.568 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.827 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:42.827 "name": "raid_bdev1", 00:13:42.827 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:42.827 "strip_size_kb": 64, 00:13:42.827 "state": "online", 00:13:42.827 "raid_level": "raid5f", 00:13:42.827 "superblock": true, 00:13:42.827 "num_base_bdevs": 3, 00:13:42.827 "num_base_bdevs_discovered": 2, 00:13:42.827 "num_base_bdevs_operational": 2, 00:13:42.827 "base_bdevs_list": [ 00:13:42.827 { 00:13:42.827 "name": null, 00:13:42.827 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.827 "is_configured": false, 00:13:42.827 "data_offset": 0, 00:13:42.827 "data_size": 63488 00:13:42.827 }, 00:13:42.827 { 00:13:42.827 "name": "BaseBdev2", 00:13:42.827 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:42.827 "is_configured": true, 00:13:42.827 "data_offset": 2048, 00:13:42.827 "data_size": 63488 00:13:42.827 }, 00:13:42.827 { 00:13:42.827 "name": "BaseBdev3", 00:13:42.827 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:42.827 "is_configured": true, 00:13:42.827 "data_offset": 2048, 00:13:42.827 "data_size": 63488 00:13:42.827 } 00:13:42.827 ] 00:13:42.827 }' 00:13:42.827 05:41:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:42.828 05:41:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:42.828 05:41:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:42.828 05:41:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:42.828 05:41:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:42.828 05:41:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.828 05:41:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.828 [2024-12-07 05:41:16.041032] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:42.828 [2024-12-07 05:41:16.045542] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000256a0 00:13:42.828 05:41:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.828 05:41:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:42.828 [2024-12-07 05:41:16.047737] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:43.768 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:43.768 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:43.768 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:43.768 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:43.768 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:43.768 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.768 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.768 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.768 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.768 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.768 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:43.768 "name": "raid_bdev1", 00:13:43.769 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:43.769 "strip_size_kb": 64, 00:13:43.769 "state": "online", 00:13:43.769 "raid_level": "raid5f", 00:13:43.769 "superblock": true, 00:13:43.769 "num_base_bdevs": 3, 00:13:43.769 "num_base_bdevs_discovered": 3, 00:13:43.769 "num_base_bdevs_operational": 3, 00:13:43.769 "process": { 00:13:43.769 "type": "rebuild", 00:13:43.769 "target": "spare", 00:13:43.769 "progress": { 00:13:43.769 "blocks": 20480, 00:13:43.769 "percent": 16 00:13:43.769 } 00:13:43.769 }, 00:13:43.769 "base_bdevs_list": [ 00:13:43.769 { 00:13:43.769 "name": "spare", 00:13:43.769 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:43.769 "is_configured": true, 00:13:43.769 "data_offset": 2048, 00:13:43.769 "data_size": 63488 00:13:43.769 }, 00:13:43.769 { 00:13:43.769 "name": "BaseBdev2", 00:13:43.769 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:43.769 "is_configured": true, 00:13:43.769 "data_offset": 2048, 00:13:43.769 "data_size": 63488 00:13:43.769 }, 00:13:43.769 { 00:13:43.769 "name": "BaseBdev3", 00:13:43.769 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:43.769 "is_configured": true, 00:13:43.769 "data_offset": 2048, 00:13:43.769 "data_size": 63488 00:13:43.769 } 00:13:43.769 ] 00:13:43.769 }' 00:13:43.769 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:44.028 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=454 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.028 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:44.028 "name": "raid_bdev1", 00:13:44.028 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:44.028 "strip_size_kb": 64, 00:13:44.028 "state": "online", 00:13:44.029 "raid_level": "raid5f", 00:13:44.029 "superblock": true, 00:13:44.029 "num_base_bdevs": 3, 00:13:44.029 "num_base_bdevs_discovered": 3, 00:13:44.029 "num_base_bdevs_operational": 3, 00:13:44.029 "process": { 00:13:44.029 "type": "rebuild", 00:13:44.029 "target": "spare", 00:13:44.029 "progress": { 00:13:44.029 "blocks": 22528, 00:13:44.029 "percent": 17 00:13:44.029 } 00:13:44.029 }, 00:13:44.029 "base_bdevs_list": [ 00:13:44.029 { 00:13:44.029 "name": "spare", 00:13:44.029 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:44.029 "is_configured": true, 00:13:44.029 "data_offset": 2048, 00:13:44.029 "data_size": 63488 00:13:44.029 }, 00:13:44.029 { 00:13:44.029 "name": "BaseBdev2", 00:13:44.029 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:44.029 "is_configured": true, 00:13:44.029 "data_offset": 2048, 00:13:44.029 "data_size": 63488 00:13:44.029 }, 00:13:44.029 { 00:13:44.029 "name": "BaseBdev3", 00:13:44.029 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:44.029 "is_configured": true, 00:13:44.029 "data_offset": 2048, 00:13:44.029 "data_size": 63488 00:13:44.029 } 00:13:44.029 ] 00:13:44.029 }' 00:13:44.029 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:44.029 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:44.029 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:44.029 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:44.029 05:41:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.412 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:45.412 "name": "raid_bdev1", 00:13:45.412 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:45.412 "strip_size_kb": 64, 00:13:45.412 "state": "online", 00:13:45.412 "raid_level": "raid5f", 00:13:45.412 "superblock": true, 00:13:45.412 "num_base_bdevs": 3, 00:13:45.412 "num_base_bdevs_discovered": 3, 00:13:45.412 "num_base_bdevs_operational": 3, 00:13:45.412 "process": { 00:13:45.412 "type": "rebuild", 00:13:45.412 "target": "spare", 00:13:45.412 "progress": { 00:13:45.412 "blocks": 47104, 00:13:45.412 "percent": 37 00:13:45.412 } 00:13:45.412 }, 00:13:45.412 "base_bdevs_list": [ 00:13:45.412 { 00:13:45.412 "name": "spare", 00:13:45.412 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:45.412 "is_configured": true, 00:13:45.412 "data_offset": 2048, 00:13:45.412 "data_size": 63488 00:13:45.412 }, 00:13:45.412 { 00:13:45.412 "name": "BaseBdev2", 00:13:45.412 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:45.412 "is_configured": true, 00:13:45.413 "data_offset": 2048, 00:13:45.413 "data_size": 63488 00:13:45.413 }, 00:13:45.413 { 00:13:45.413 "name": "BaseBdev3", 00:13:45.413 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:45.413 "is_configured": true, 00:13:45.413 "data_offset": 2048, 00:13:45.413 "data_size": 63488 00:13:45.413 } 00:13:45.413 ] 00:13:45.413 }' 00:13:45.413 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:45.413 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:45.413 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:45.413 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:45.413 05:41:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.351 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:46.352 "name": "raid_bdev1", 00:13:46.352 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:46.352 "strip_size_kb": 64, 00:13:46.352 "state": "online", 00:13:46.352 "raid_level": "raid5f", 00:13:46.352 "superblock": true, 00:13:46.352 "num_base_bdevs": 3, 00:13:46.352 "num_base_bdevs_discovered": 3, 00:13:46.352 "num_base_bdevs_operational": 3, 00:13:46.352 "process": { 00:13:46.352 "type": "rebuild", 00:13:46.352 "target": "spare", 00:13:46.352 "progress": { 00:13:46.352 "blocks": 69632, 00:13:46.352 "percent": 54 00:13:46.352 } 00:13:46.352 }, 00:13:46.352 "base_bdevs_list": [ 00:13:46.352 { 00:13:46.352 "name": "spare", 00:13:46.352 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:46.352 "is_configured": true, 00:13:46.352 "data_offset": 2048, 00:13:46.352 "data_size": 63488 00:13:46.352 }, 00:13:46.352 { 00:13:46.352 "name": "BaseBdev2", 00:13:46.352 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:46.352 "is_configured": true, 00:13:46.352 "data_offset": 2048, 00:13:46.352 "data_size": 63488 00:13:46.352 }, 00:13:46.352 { 00:13:46.352 "name": "BaseBdev3", 00:13:46.352 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:46.352 "is_configured": true, 00:13:46.352 "data_offset": 2048, 00:13:46.352 "data_size": 63488 00:13:46.352 } 00:13:46.352 ] 00:13:46.352 }' 00:13:46.352 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:46.352 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:46.352 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:46.352 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:46.352 05:41:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.731 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:47.731 "name": "raid_bdev1", 00:13:47.731 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:47.731 "strip_size_kb": 64, 00:13:47.731 "state": "online", 00:13:47.731 "raid_level": "raid5f", 00:13:47.731 "superblock": true, 00:13:47.731 "num_base_bdevs": 3, 00:13:47.731 "num_base_bdevs_discovered": 3, 00:13:47.731 "num_base_bdevs_operational": 3, 00:13:47.731 "process": { 00:13:47.731 "type": "rebuild", 00:13:47.731 "target": "spare", 00:13:47.731 "progress": { 00:13:47.731 "blocks": 92160, 00:13:47.731 "percent": 72 00:13:47.731 } 00:13:47.731 }, 00:13:47.731 "base_bdevs_list": [ 00:13:47.731 { 00:13:47.731 "name": "spare", 00:13:47.731 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:47.731 "is_configured": true, 00:13:47.731 "data_offset": 2048, 00:13:47.731 "data_size": 63488 00:13:47.731 }, 00:13:47.731 { 00:13:47.731 "name": "BaseBdev2", 00:13:47.731 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:47.731 "is_configured": true, 00:13:47.731 "data_offset": 2048, 00:13:47.731 "data_size": 63488 00:13:47.731 }, 00:13:47.731 { 00:13:47.731 "name": "BaseBdev3", 00:13:47.731 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:47.732 "is_configured": true, 00:13:47.732 "data_offset": 2048, 00:13:47.732 "data_size": 63488 00:13:47.732 } 00:13:47.732 ] 00:13:47.732 }' 00:13:47.732 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:47.732 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:47.732 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:47.732 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:47.732 05:41:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.668 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:48.668 "name": "raid_bdev1", 00:13:48.668 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:48.668 "strip_size_kb": 64, 00:13:48.668 "state": "online", 00:13:48.668 "raid_level": "raid5f", 00:13:48.668 "superblock": true, 00:13:48.668 "num_base_bdevs": 3, 00:13:48.668 "num_base_bdevs_discovered": 3, 00:13:48.668 "num_base_bdevs_operational": 3, 00:13:48.668 "process": { 00:13:48.668 "type": "rebuild", 00:13:48.668 "target": "spare", 00:13:48.668 "progress": { 00:13:48.668 "blocks": 116736, 00:13:48.668 "percent": 91 00:13:48.668 } 00:13:48.668 }, 00:13:48.668 "base_bdevs_list": [ 00:13:48.668 { 00:13:48.668 "name": "spare", 00:13:48.668 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:48.668 "is_configured": true, 00:13:48.668 "data_offset": 2048, 00:13:48.668 "data_size": 63488 00:13:48.668 }, 00:13:48.668 { 00:13:48.668 "name": "BaseBdev2", 00:13:48.668 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:48.668 "is_configured": true, 00:13:48.668 "data_offset": 2048, 00:13:48.668 "data_size": 63488 00:13:48.668 }, 00:13:48.668 { 00:13:48.668 "name": "BaseBdev3", 00:13:48.668 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:48.669 "is_configured": true, 00:13:48.669 "data_offset": 2048, 00:13:48.669 "data_size": 63488 00:13:48.669 } 00:13:48.669 ] 00:13:48.669 }' 00:13:48.669 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:48.669 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:48.669 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:48.669 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:48.669 05:41:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:48.928 [2024-12-07 05:41:22.282493] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:48.928 [2024-12-07 05:41:22.282555] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:48.928 [2024-12-07 05:41:22.282669] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:49.864 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:49.865 "name": "raid_bdev1", 00:13:49.865 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:49.865 "strip_size_kb": 64, 00:13:49.865 "state": "online", 00:13:49.865 "raid_level": "raid5f", 00:13:49.865 "superblock": true, 00:13:49.865 "num_base_bdevs": 3, 00:13:49.865 "num_base_bdevs_discovered": 3, 00:13:49.865 "num_base_bdevs_operational": 3, 00:13:49.865 "base_bdevs_list": [ 00:13:49.865 { 00:13:49.865 "name": "spare", 00:13:49.865 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:49.865 "is_configured": true, 00:13:49.865 "data_offset": 2048, 00:13:49.865 "data_size": 63488 00:13:49.865 }, 00:13:49.865 { 00:13:49.865 "name": "BaseBdev2", 00:13:49.865 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:49.865 "is_configured": true, 00:13:49.865 "data_offset": 2048, 00:13:49.865 "data_size": 63488 00:13:49.865 }, 00:13:49.865 { 00:13:49.865 "name": "BaseBdev3", 00:13:49.865 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:49.865 "is_configured": true, 00:13:49.865 "data_offset": 2048, 00:13:49.865 "data_size": 63488 00:13:49.865 } 00:13:49.865 ] 00:13:49.865 }' 00:13:49.865 05:41:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:49.865 "name": "raid_bdev1", 00:13:49.865 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:49.865 "strip_size_kb": 64, 00:13:49.865 "state": "online", 00:13:49.865 "raid_level": "raid5f", 00:13:49.865 "superblock": true, 00:13:49.865 "num_base_bdevs": 3, 00:13:49.865 "num_base_bdevs_discovered": 3, 00:13:49.865 "num_base_bdevs_operational": 3, 00:13:49.865 "base_bdevs_list": [ 00:13:49.865 { 00:13:49.865 "name": "spare", 00:13:49.865 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:49.865 "is_configured": true, 00:13:49.865 "data_offset": 2048, 00:13:49.865 "data_size": 63488 00:13:49.865 }, 00:13:49.865 { 00:13:49.865 "name": "BaseBdev2", 00:13:49.865 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:49.865 "is_configured": true, 00:13:49.865 "data_offset": 2048, 00:13:49.865 "data_size": 63488 00:13:49.865 }, 00:13:49.865 { 00:13:49.865 "name": "BaseBdev3", 00:13:49.865 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:49.865 "is_configured": true, 00:13:49.865 "data_offset": 2048, 00:13:49.865 "data_size": 63488 00:13:49.865 } 00:13:49.865 ] 00:13:49.865 }' 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.865 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.124 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.124 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:50.124 "name": "raid_bdev1", 00:13:50.124 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:50.124 "strip_size_kb": 64, 00:13:50.124 "state": "online", 00:13:50.124 "raid_level": "raid5f", 00:13:50.124 "superblock": true, 00:13:50.124 "num_base_bdevs": 3, 00:13:50.124 "num_base_bdevs_discovered": 3, 00:13:50.124 "num_base_bdevs_operational": 3, 00:13:50.124 "base_bdevs_list": [ 00:13:50.124 { 00:13:50.124 "name": "spare", 00:13:50.124 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:50.124 "is_configured": true, 00:13:50.124 "data_offset": 2048, 00:13:50.124 "data_size": 63488 00:13:50.124 }, 00:13:50.124 { 00:13:50.124 "name": "BaseBdev2", 00:13:50.124 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:50.124 "is_configured": true, 00:13:50.124 "data_offset": 2048, 00:13:50.124 "data_size": 63488 00:13:50.124 }, 00:13:50.124 { 00:13:50.124 "name": "BaseBdev3", 00:13:50.124 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:50.124 "is_configured": true, 00:13:50.124 "data_offset": 2048, 00:13:50.124 "data_size": 63488 00:13:50.124 } 00:13:50.124 ] 00:13:50.124 }' 00:13:50.124 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:50.124 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.382 [2024-12-07 05:41:23.657767] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:50.382 [2024-12-07 05:41:23.657875] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:50.382 [2024-12-07 05:41:23.657988] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:50.382 [2024-12-07 05:41:23.658111] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:50.382 [2024-12-07 05:41:23.658166] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:50.382 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:50.383 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:50.383 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:50.383 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:50.383 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:50.383 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:50.383 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:50.383 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:50.383 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:50.641 /dev/nbd0 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:50.641 1+0 records in 00:13:50.641 1+0 records out 00:13:50.641 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000556412 s, 7.4 MB/s 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:50.641 05:41:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:50.900 /dev/nbd1 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:50.900 1+0 records in 00:13:50.900 1+0 records out 00:13:50.900 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000345994 s, 11.8 MB/s 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:50.900 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:51.159 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.419 [2024-12-07 05:41:24.743761] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:51.419 [2024-12-07 05:41:24.743813] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:51.419 [2024-12-07 05:41:24.743834] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:13:51.419 [2024-12-07 05:41:24.743843] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:51.419 [2024-12-07 05:41:24.745993] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:51.419 [2024-12-07 05:41:24.746016] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:51.419 [2024-12-07 05:41:24.746098] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:51.419 [2024-12-07 05:41:24.746144] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:51.419 [2024-12-07 05:41:24.746268] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:51.419 [2024-12-07 05:41:24.746357] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:51.419 spare 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.419 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.679 [2024-12-07 05:41:24.846256] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:13:51.679 [2024-12-07 05:41:24.846283] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:51.679 [2024-12-07 05:41:24.846568] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043d50 00:13:51.679 [2024-12-07 05:41:24.847011] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:13:51.679 [2024-12-07 05:41:24.847028] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:13:51.679 [2024-12-07 05:41:24.847194] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:51.679 "name": "raid_bdev1", 00:13:51.679 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:51.679 "strip_size_kb": 64, 00:13:51.679 "state": "online", 00:13:51.679 "raid_level": "raid5f", 00:13:51.679 "superblock": true, 00:13:51.679 "num_base_bdevs": 3, 00:13:51.679 "num_base_bdevs_discovered": 3, 00:13:51.679 "num_base_bdevs_operational": 3, 00:13:51.679 "base_bdevs_list": [ 00:13:51.679 { 00:13:51.679 "name": "spare", 00:13:51.679 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:51.679 "is_configured": true, 00:13:51.679 "data_offset": 2048, 00:13:51.679 "data_size": 63488 00:13:51.679 }, 00:13:51.679 { 00:13:51.679 "name": "BaseBdev2", 00:13:51.679 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:51.679 "is_configured": true, 00:13:51.679 "data_offset": 2048, 00:13:51.679 "data_size": 63488 00:13:51.679 }, 00:13:51.679 { 00:13:51.679 "name": "BaseBdev3", 00:13:51.679 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:51.679 "is_configured": true, 00:13:51.679 "data_offset": 2048, 00:13:51.679 "data_size": 63488 00:13:51.679 } 00:13:51.679 ] 00:13:51.679 }' 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:51.679 05:41:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.939 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:51.939 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:51.939 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:51.939 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:51.939 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:51.939 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:51.939 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.939 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.940 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.940 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.940 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:51.940 "name": "raid_bdev1", 00:13:51.940 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:51.940 "strip_size_kb": 64, 00:13:51.940 "state": "online", 00:13:51.940 "raid_level": "raid5f", 00:13:51.940 "superblock": true, 00:13:51.940 "num_base_bdevs": 3, 00:13:51.940 "num_base_bdevs_discovered": 3, 00:13:51.940 "num_base_bdevs_operational": 3, 00:13:51.940 "base_bdevs_list": [ 00:13:51.940 { 00:13:51.940 "name": "spare", 00:13:51.940 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:51.940 "is_configured": true, 00:13:51.940 "data_offset": 2048, 00:13:51.940 "data_size": 63488 00:13:51.940 }, 00:13:51.940 { 00:13:51.940 "name": "BaseBdev2", 00:13:51.940 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:51.940 "is_configured": true, 00:13:51.940 "data_offset": 2048, 00:13:51.940 "data_size": 63488 00:13:51.940 }, 00:13:51.940 { 00:13:51.940 "name": "BaseBdev3", 00:13:51.940 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:51.940 "is_configured": true, 00:13:51.940 "data_offset": 2048, 00:13:51.940 "data_size": 63488 00:13:51.940 } 00:13:51.940 ] 00:13:51.940 }' 00:13:51.940 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.199 [2024-12-07 05:41:25.451353] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:52.199 "name": "raid_bdev1", 00:13:52.199 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:52.199 "strip_size_kb": 64, 00:13:52.199 "state": "online", 00:13:52.199 "raid_level": "raid5f", 00:13:52.199 "superblock": true, 00:13:52.199 "num_base_bdevs": 3, 00:13:52.199 "num_base_bdevs_discovered": 2, 00:13:52.199 "num_base_bdevs_operational": 2, 00:13:52.199 "base_bdevs_list": [ 00:13:52.199 { 00:13:52.199 "name": null, 00:13:52.199 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:52.199 "is_configured": false, 00:13:52.199 "data_offset": 0, 00:13:52.199 "data_size": 63488 00:13:52.199 }, 00:13:52.199 { 00:13:52.199 "name": "BaseBdev2", 00:13:52.199 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:52.199 "is_configured": true, 00:13:52.199 "data_offset": 2048, 00:13:52.199 "data_size": 63488 00:13:52.199 }, 00:13:52.199 { 00:13:52.199 "name": "BaseBdev3", 00:13:52.199 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:52.199 "is_configured": true, 00:13:52.199 "data_offset": 2048, 00:13:52.199 "data_size": 63488 00:13:52.199 } 00:13:52.199 ] 00:13:52.199 }' 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:52.199 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.768 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:52.768 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.768 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.768 [2024-12-07 05:41:25.894646] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:52.768 [2024-12-07 05:41:25.894879] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:13:52.768 [2024-12-07 05:41:25.894941] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:52.768 [2024-12-07 05:41:25.895028] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:52.768 [2024-12-07 05:41:25.899411] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043e20 00:13:52.768 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.768 05:41:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:52.768 [2024-12-07 05:41:25.901575] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:53.707 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:53.707 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:53.707 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:53.707 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:53.707 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:53.707 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.707 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.707 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.707 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.707 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.707 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:53.707 "name": "raid_bdev1", 00:13:53.707 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:53.707 "strip_size_kb": 64, 00:13:53.707 "state": "online", 00:13:53.707 "raid_level": "raid5f", 00:13:53.707 "superblock": true, 00:13:53.707 "num_base_bdevs": 3, 00:13:53.707 "num_base_bdevs_discovered": 3, 00:13:53.707 "num_base_bdevs_operational": 3, 00:13:53.707 "process": { 00:13:53.707 "type": "rebuild", 00:13:53.707 "target": "spare", 00:13:53.707 "progress": { 00:13:53.707 "blocks": 20480, 00:13:53.707 "percent": 16 00:13:53.707 } 00:13:53.707 }, 00:13:53.707 "base_bdevs_list": [ 00:13:53.707 { 00:13:53.707 "name": "spare", 00:13:53.707 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:53.707 "is_configured": true, 00:13:53.708 "data_offset": 2048, 00:13:53.708 "data_size": 63488 00:13:53.708 }, 00:13:53.708 { 00:13:53.708 "name": "BaseBdev2", 00:13:53.708 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:53.708 "is_configured": true, 00:13:53.708 "data_offset": 2048, 00:13:53.708 "data_size": 63488 00:13:53.708 }, 00:13:53.708 { 00:13:53.708 "name": "BaseBdev3", 00:13:53.708 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:53.708 "is_configured": true, 00:13:53.708 "data_offset": 2048, 00:13:53.708 "data_size": 63488 00:13:53.708 } 00:13:53.708 ] 00:13:53.708 }' 00:13:53.708 05:41:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:53.708 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:53.708 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:53.708 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:53.708 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:53.708 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.708 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.708 [2024-12-07 05:41:27.061710] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:53.968 [2024-12-07 05:41:27.108620] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:53.968 [2024-12-07 05:41:27.108682] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:53.968 [2024-12-07 05:41:27.108700] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:53.968 [2024-12-07 05:41:27.108707] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:53.968 "name": "raid_bdev1", 00:13:53.968 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:53.968 "strip_size_kb": 64, 00:13:53.968 "state": "online", 00:13:53.968 "raid_level": "raid5f", 00:13:53.968 "superblock": true, 00:13:53.968 "num_base_bdevs": 3, 00:13:53.968 "num_base_bdevs_discovered": 2, 00:13:53.968 "num_base_bdevs_operational": 2, 00:13:53.968 "base_bdevs_list": [ 00:13:53.968 { 00:13:53.968 "name": null, 00:13:53.968 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:53.968 "is_configured": false, 00:13:53.968 "data_offset": 0, 00:13:53.968 "data_size": 63488 00:13:53.968 }, 00:13:53.968 { 00:13:53.968 "name": "BaseBdev2", 00:13:53.968 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:53.968 "is_configured": true, 00:13:53.968 "data_offset": 2048, 00:13:53.968 "data_size": 63488 00:13:53.968 }, 00:13:53.968 { 00:13:53.968 "name": "BaseBdev3", 00:13:53.968 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:53.968 "is_configured": true, 00:13:53.968 "data_offset": 2048, 00:13:53.968 "data_size": 63488 00:13:53.968 } 00:13:53.968 ] 00:13:53.968 }' 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:53.968 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:54.228 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:54.228 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.228 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:54.228 [2024-12-07 05:41:27.533252] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:54.228 [2024-12-07 05:41:27.533353] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:54.228 [2024-12-07 05:41:27.533409] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:13:54.228 [2024-12-07 05:41:27.533437] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:54.228 [2024-12-07 05:41:27.533896] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:54.228 [2024-12-07 05:41:27.533960] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:54.228 [2024-12-07 05:41:27.534067] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:54.228 [2024-12-07 05:41:27.534108] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:13:54.228 [2024-12-07 05:41:27.534152] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:54.228 [2024-12-07 05:41:27.534228] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:54.228 [2024-12-07 05:41:27.538245] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043ef0 00:13:54.228 spare 00:13:54.228 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.228 05:41:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:54.228 [2024-12-07 05:41:27.540374] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:55.612 "name": "raid_bdev1", 00:13:55.612 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:55.612 "strip_size_kb": 64, 00:13:55.612 "state": "online", 00:13:55.612 "raid_level": "raid5f", 00:13:55.612 "superblock": true, 00:13:55.612 "num_base_bdevs": 3, 00:13:55.612 "num_base_bdevs_discovered": 3, 00:13:55.612 "num_base_bdevs_operational": 3, 00:13:55.612 "process": { 00:13:55.612 "type": "rebuild", 00:13:55.612 "target": "spare", 00:13:55.612 "progress": { 00:13:55.612 "blocks": 20480, 00:13:55.612 "percent": 16 00:13:55.612 } 00:13:55.612 }, 00:13:55.612 "base_bdevs_list": [ 00:13:55.612 { 00:13:55.612 "name": "spare", 00:13:55.612 "uuid": "6c3e40c0-0ff8-52e4-8580-ed5499c8a70b", 00:13:55.612 "is_configured": true, 00:13:55.612 "data_offset": 2048, 00:13:55.612 "data_size": 63488 00:13:55.612 }, 00:13:55.612 { 00:13:55.612 "name": "BaseBdev2", 00:13:55.612 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:55.612 "is_configured": true, 00:13:55.612 "data_offset": 2048, 00:13:55.612 "data_size": 63488 00:13:55.612 }, 00:13:55.612 { 00:13:55.612 "name": "BaseBdev3", 00:13:55.612 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:55.612 "is_configured": true, 00:13:55.612 "data_offset": 2048, 00:13:55.612 "data_size": 63488 00:13:55.612 } 00:13:55.612 ] 00:13:55.612 }' 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:55.612 [2024-12-07 05:41:28.700757] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:55.612 [2024-12-07 05:41:28.747320] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:55.612 [2024-12-07 05:41:28.747377] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:55.612 [2024-12-07 05:41:28.747392] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:55.612 [2024-12-07 05:41:28.747403] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:55.612 "name": "raid_bdev1", 00:13:55.612 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:55.612 "strip_size_kb": 64, 00:13:55.612 "state": "online", 00:13:55.612 "raid_level": "raid5f", 00:13:55.612 "superblock": true, 00:13:55.612 "num_base_bdevs": 3, 00:13:55.612 "num_base_bdevs_discovered": 2, 00:13:55.612 "num_base_bdevs_operational": 2, 00:13:55.612 "base_bdevs_list": [ 00:13:55.612 { 00:13:55.612 "name": null, 00:13:55.612 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:55.612 "is_configured": false, 00:13:55.612 "data_offset": 0, 00:13:55.612 "data_size": 63488 00:13:55.612 }, 00:13:55.612 { 00:13:55.612 "name": "BaseBdev2", 00:13:55.612 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:55.612 "is_configured": true, 00:13:55.612 "data_offset": 2048, 00:13:55.612 "data_size": 63488 00:13:55.612 }, 00:13:55.612 { 00:13:55.612 "name": "BaseBdev3", 00:13:55.612 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:55.612 "is_configured": true, 00:13:55.612 "data_offset": 2048, 00:13:55.612 "data_size": 63488 00:13:55.612 } 00:13:55.612 ] 00:13:55.612 }' 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:55.612 05:41:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:55.872 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:55.872 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:55.872 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:55.872 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:55.872 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:55.872 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:55.872 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.872 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:55.872 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:55.872 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:55.872 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:55.872 "name": "raid_bdev1", 00:13:55.872 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:55.872 "strip_size_kb": 64, 00:13:55.872 "state": "online", 00:13:55.872 "raid_level": "raid5f", 00:13:55.872 "superblock": true, 00:13:55.873 "num_base_bdevs": 3, 00:13:55.873 "num_base_bdevs_discovered": 2, 00:13:55.873 "num_base_bdevs_operational": 2, 00:13:55.873 "base_bdevs_list": [ 00:13:55.873 { 00:13:55.873 "name": null, 00:13:55.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:55.873 "is_configured": false, 00:13:55.873 "data_offset": 0, 00:13:55.873 "data_size": 63488 00:13:55.873 }, 00:13:55.873 { 00:13:55.873 "name": "BaseBdev2", 00:13:55.873 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:55.873 "is_configured": true, 00:13:55.873 "data_offset": 2048, 00:13:55.873 "data_size": 63488 00:13:55.873 }, 00:13:55.873 { 00:13:55.873 "name": "BaseBdev3", 00:13:55.873 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:55.873 "is_configured": true, 00:13:55.873 "data_offset": 2048, 00:13:55.873 "data_size": 63488 00:13:55.873 } 00:13:55.873 ] 00:13:55.873 }' 00:13:55.873 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.132 [2024-12-07 05:41:29.327725] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:56.132 [2024-12-07 05:41:29.327827] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:56.132 [2024-12-07 05:41:29.327851] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:13:56.132 [2024-12-07 05:41:29.327862] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:56.132 [2024-12-07 05:41:29.328246] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:56.132 [2024-12-07 05:41:29.328266] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:56.132 [2024-12-07 05:41:29.328332] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:56.132 [2024-12-07 05:41:29.328348] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:13:56.132 [2024-12-07 05:41:29.328356] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:56.132 [2024-12-07 05:41:29.328368] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:56.132 BaseBdev1 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.132 05:41:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:57.114 "name": "raid_bdev1", 00:13:57.114 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:57.114 "strip_size_kb": 64, 00:13:57.114 "state": "online", 00:13:57.114 "raid_level": "raid5f", 00:13:57.114 "superblock": true, 00:13:57.114 "num_base_bdevs": 3, 00:13:57.114 "num_base_bdevs_discovered": 2, 00:13:57.114 "num_base_bdevs_operational": 2, 00:13:57.114 "base_bdevs_list": [ 00:13:57.114 { 00:13:57.114 "name": null, 00:13:57.114 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:57.114 "is_configured": false, 00:13:57.114 "data_offset": 0, 00:13:57.114 "data_size": 63488 00:13:57.114 }, 00:13:57.114 { 00:13:57.114 "name": "BaseBdev2", 00:13:57.114 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:57.114 "is_configured": true, 00:13:57.114 "data_offset": 2048, 00:13:57.114 "data_size": 63488 00:13:57.114 }, 00:13:57.114 { 00:13:57.114 "name": "BaseBdev3", 00:13:57.114 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:57.114 "is_configured": true, 00:13:57.114 "data_offset": 2048, 00:13:57.114 "data_size": 63488 00:13:57.114 } 00:13:57.114 ] 00:13:57.114 }' 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:57.114 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.392 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:57.392 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:57.392 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:57.392 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:57.392 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:57.671 "name": "raid_bdev1", 00:13:57.671 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:57.671 "strip_size_kb": 64, 00:13:57.671 "state": "online", 00:13:57.671 "raid_level": "raid5f", 00:13:57.671 "superblock": true, 00:13:57.671 "num_base_bdevs": 3, 00:13:57.671 "num_base_bdevs_discovered": 2, 00:13:57.671 "num_base_bdevs_operational": 2, 00:13:57.671 "base_bdevs_list": [ 00:13:57.671 { 00:13:57.671 "name": null, 00:13:57.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:57.671 "is_configured": false, 00:13:57.671 "data_offset": 0, 00:13:57.671 "data_size": 63488 00:13:57.671 }, 00:13:57.671 { 00:13:57.671 "name": "BaseBdev2", 00:13:57.671 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:57.671 "is_configured": true, 00:13:57.671 "data_offset": 2048, 00:13:57.671 "data_size": 63488 00:13:57.671 }, 00:13:57.671 { 00:13:57.671 "name": "BaseBdev3", 00:13:57.671 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:57.671 "is_configured": true, 00:13:57.671 "data_offset": 2048, 00:13:57.671 "data_size": 63488 00:13:57.671 } 00:13:57.671 ] 00:13:57.671 }' 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.671 [2024-12-07 05:41:30.861146] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:57.671 [2024-12-07 05:41:30.861298] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:13:57.671 [2024-12-07 05:41:30.861312] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:57.671 request: 00:13:57.671 { 00:13:57.671 "base_bdev": "BaseBdev1", 00:13:57.671 "raid_bdev": "raid_bdev1", 00:13:57.671 "method": "bdev_raid_add_base_bdev", 00:13:57.671 "req_id": 1 00:13:57.671 } 00:13:57.671 Got JSON-RPC error response 00:13:57.671 response: 00:13:57.671 { 00:13:57.671 "code": -22, 00:13:57.671 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:57.671 } 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:57.671 05:41:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:58.610 "name": "raid_bdev1", 00:13:58.610 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:58.610 "strip_size_kb": 64, 00:13:58.610 "state": "online", 00:13:58.610 "raid_level": "raid5f", 00:13:58.610 "superblock": true, 00:13:58.610 "num_base_bdevs": 3, 00:13:58.610 "num_base_bdevs_discovered": 2, 00:13:58.610 "num_base_bdevs_operational": 2, 00:13:58.610 "base_bdevs_list": [ 00:13:58.610 { 00:13:58.610 "name": null, 00:13:58.610 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:58.610 "is_configured": false, 00:13:58.610 "data_offset": 0, 00:13:58.610 "data_size": 63488 00:13:58.610 }, 00:13:58.610 { 00:13:58.610 "name": "BaseBdev2", 00:13:58.610 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:58.610 "is_configured": true, 00:13:58.610 "data_offset": 2048, 00:13:58.610 "data_size": 63488 00:13:58.610 }, 00:13:58.610 { 00:13:58.610 "name": "BaseBdev3", 00:13:58.610 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:58.610 "is_configured": true, 00:13:58.610 "data_offset": 2048, 00:13:58.610 "data_size": 63488 00:13:58.610 } 00:13:58.610 ] 00:13:58.610 }' 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:58.610 05:41:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:59.180 "name": "raid_bdev1", 00:13:59.180 "uuid": "1bdf8e95-064d-4d34-97b5-acbd5acc1712", 00:13:59.180 "strip_size_kb": 64, 00:13:59.180 "state": "online", 00:13:59.180 "raid_level": "raid5f", 00:13:59.180 "superblock": true, 00:13:59.180 "num_base_bdevs": 3, 00:13:59.180 "num_base_bdevs_discovered": 2, 00:13:59.180 "num_base_bdevs_operational": 2, 00:13:59.180 "base_bdevs_list": [ 00:13:59.180 { 00:13:59.180 "name": null, 00:13:59.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:59.180 "is_configured": false, 00:13:59.180 "data_offset": 0, 00:13:59.180 "data_size": 63488 00:13:59.180 }, 00:13:59.180 { 00:13:59.180 "name": "BaseBdev2", 00:13:59.180 "uuid": "707fc60d-823c-5431-a74d-8cef0267e01a", 00:13:59.180 "is_configured": true, 00:13:59.180 "data_offset": 2048, 00:13:59.180 "data_size": 63488 00:13:59.180 }, 00:13:59.180 { 00:13:59.180 "name": "BaseBdev3", 00:13:59.180 "uuid": "552128a4-1c65-56d9-8e1d-700737aa587f", 00:13:59.180 "is_configured": true, 00:13:59.180 "data_offset": 2048, 00:13:59.180 "data_size": 63488 00:13:59.180 } 00:13:59.180 ] 00:13:59.180 }' 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 92118 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 92118 ']' 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 92118 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 92118 00:13:59.180 killing process with pid 92118 00:13:59.180 Received shutdown signal, test time was about 60.000000 seconds 00:13:59.180 00:13:59.180 Latency(us) 00:13:59.180 [2024-12-07T05:41:32.548Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:59.180 [2024-12-07T05:41:32.548Z] =================================================================================================================== 00:13:59.180 [2024-12-07T05:41:32.548Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 92118' 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 92118 00:13:59.180 [2024-12-07 05:41:32.510230] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:59.180 [2024-12-07 05:41:32.510356] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:59.180 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 92118 00:13:59.180 [2024-12-07 05:41:32.510430] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:59.180 [2024-12-07 05:41:32.510440] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:13:59.440 [2024-12-07 05:41:32.551322] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:59.440 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:13:59.440 00:13:59.440 real 0m21.273s 00:13:59.440 user 0m27.619s 00:13:59.440 sys 0m2.707s 00:13:59.440 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:59.440 05:41:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.440 ************************************ 00:13:59.440 END TEST raid5f_rebuild_test_sb 00:13:59.440 ************************************ 00:13:59.440 05:41:32 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:13:59.440 05:41:32 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:13:59.440 05:41:32 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:13:59.440 05:41:32 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:59.440 05:41:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:59.700 ************************************ 00:13:59.700 START TEST raid5f_state_function_test 00:13:59.700 ************************************ 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 false 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:59.700 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=92846 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 92846' 00:13:59.701 Process raid pid: 92846 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 92846 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 92846 ']' 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:59.701 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:59.701 05:41:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:59.701 [2024-12-07 05:41:32.910223] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:13:59.701 [2024-12-07 05:41:32.910441] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:59.701 [2024-12-07 05:41:33.065849] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:59.960 [2024-12-07 05:41:33.090319] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:59.960 [2024-12-07 05:41:33.132639] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:59.960 [2024-12-07 05:41:33.132768] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:00.537 [2024-12-07 05:41:33.743235] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:00.537 [2024-12-07 05:41:33.743295] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:00.537 [2024-12-07 05:41:33.743305] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:00.537 [2024-12-07 05:41:33.743315] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:00.537 [2024-12-07 05:41:33.743322] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:00.537 [2024-12-07 05:41:33.743334] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:00.537 [2024-12-07 05:41:33.743340] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:00.537 [2024-12-07 05:41:33.743349] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:00.537 "name": "Existed_Raid", 00:14:00.537 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.537 "strip_size_kb": 64, 00:14:00.537 "state": "configuring", 00:14:00.537 "raid_level": "raid5f", 00:14:00.537 "superblock": false, 00:14:00.537 "num_base_bdevs": 4, 00:14:00.537 "num_base_bdevs_discovered": 0, 00:14:00.537 "num_base_bdevs_operational": 4, 00:14:00.537 "base_bdevs_list": [ 00:14:00.537 { 00:14:00.537 "name": "BaseBdev1", 00:14:00.537 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.537 "is_configured": false, 00:14:00.537 "data_offset": 0, 00:14:00.537 "data_size": 0 00:14:00.537 }, 00:14:00.537 { 00:14:00.537 "name": "BaseBdev2", 00:14:00.537 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.537 "is_configured": false, 00:14:00.537 "data_offset": 0, 00:14:00.537 "data_size": 0 00:14:00.537 }, 00:14:00.537 { 00:14:00.537 "name": "BaseBdev3", 00:14:00.537 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.537 "is_configured": false, 00:14:00.537 "data_offset": 0, 00:14:00.537 "data_size": 0 00:14:00.537 }, 00:14:00.537 { 00:14:00.537 "name": "BaseBdev4", 00:14:00.537 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.537 "is_configured": false, 00:14:00.537 "data_offset": 0, 00:14:00.537 "data_size": 0 00:14:00.537 } 00:14:00.537 ] 00:14:00.537 }' 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:00.537 05:41:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:00.797 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:00.797 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.058 [2024-12-07 05:41:34.170424] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:01.058 [2024-12-07 05:41:34.170508] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.058 [2024-12-07 05:41:34.182417] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:01.058 [2024-12-07 05:41:34.182519] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:01.058 [2024-12-07 05:41:34.182546] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:01.058 [2024-12-07 05:41:34.182568] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:01.058 [2024-12-07 05:41:34.182586] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:01.058 [2024-12-07 05:41:34.182606] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:01.058 [2024-12-07 05:41:34.182631] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:01.058 [2024-12-07 05:41:34.182652] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.058 [2024-12-07 05:41:34.203149] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:01.058 BaseBdev1 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.058 [ 00:14:01.058 { 00:14:01.058 "name": "BaseBdev1", 00:14:01.058 "aliases": [ 00:14:01.058 "ef226877-8793-43a3-a398-13a79349d4d7" 00:14:01.058 ], 00:14:01.058 "product_name": "Malloc disk", 00:14:01.058 "block_size": 512, 00:14:01.058 "num_blocks": 65536, 00:14:01.058 "uuid": "ef226877-8793-43a3-a398-13a79349d4d7", 00:14:01.058 "assigned_rate_limits": { 00:14:01.058 "rw_ios_per_sec": 0, 00:14:01.058 "rw_mbytes_per_sec": 0, 00:14:01.058 "r_mbytes_per_sec": 0, 00:14:01.058 "w_mbytes_per_sec": 0 00:14:01.058 }, 00:14:01.058 "claimed": true, 00:14:01.058 "claim_type": "exclusive_write", 00:14:01.058 "zoned": false, 00:14:01.058 "supported_io_types": { 00:14:01.058 "read": true, 00:14:01.058 "write": true, 00:14:01.058 "unmap": true, 00:14:01.058 "flush": true, 00:14:01.058 "reset": true, 00:14:01.058 "nvme_admin": false, 00:14:01.058 "nvme_io": false, 00:14:01.058 "nvme_io_md": false, 00:14:01.058 "write_zeroes": true, 00:14:01.058 "zcopy": true, 00:14:01.058 "get_zone_info": false, 00:14:01.058 "zone_management": false, 00:14:01.058 "zone_append": false, 00:14:01.058 "compare": false, 00:14:01.058 "compare_and_write": false, 00:14:01.058 "abort": true, 00:14:01.058 "seek_hole": false, 00:14:01.058 "seek_data": false, 00:14:01.058 "copy": true, 00:14:01.058 "nvme_iov_md": false 00:14:01.058 }, 00:14:01.058 "memory_domains": [ 00:14:01.058 { 00:14:01.058 "dma_device_id": "system", 00:14:01.058 "dma_device_type": 1 00:14:01.058 }, 00:14:01.058 { 00:14:01.058 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:01.058 "dma_device_type": 2 00:14:01.058 } 00:14:01.058 ], 00:14:01.058 "driver_specific": {} 00:14:01.058 } 00:14:01.058 ] 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:01.058 "name": "Existed_Raid", 00:14:01.058 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.058 "strip_size_kb": 64, 00:14:01.058 "state": "configuring", 00:14:01.058 "raid_level": "raid5f", 00:14:01.058 "superblock": false, 00:14:01.058 "num_base_bdevs": 4, 00:14:01.058 "num_base_bdevs_discovered": 1, 00:14:01.058 "num_base_bdevs_operational": 4, 00:14:01.058 "base_bdevs_list": [ 00:14:01.058 { 00:14:01.058 "name": "BaseBdev1", 00:14:01.058 "uuid": "ef226877-8793-43a3-a398-13a79349d4d7", 00:14:01.058 "is_configured": true, 00:14:01.058 "data_offset": 0, 00:14:01.058 "data_size": 65536 00:14:01.058 }, 00:14:01.058 { 00:14:01.058 "name": "BaseBdev2", 00:14:01.058 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.058 "is_configured": false, 00:14:01.058 "data_offset": 0, 00:14:01.058 "data_size": 0 00:14:01.058 }, 00:14:01.058 { 00:14:01.058 "name": "BaseBdev3", 00:14:01.058 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.058 "is_configured": false, 00:14:01.058 "data_offset": 0, 00:14:01.058 "data_size": 0 00:14:01.058 }, 00:14:01.058 { 00:14:01.058 "name": "BaseBdev4", 00:14:01.058 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.058 "is_configured": false, 00:14:01.058 "data_offset": 0, 00:14:01.058 "data_size": 0 00:14:01.058 } 00:14:01.058 ] 00:14:01.058 }' 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:01.058 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.319 [2024-12-07 05:41:34.614494] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:01.319 [2024-12-07 05:41:34.614541] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.319 [2024-12-07 05:41:34.626535] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:01.319 [2024-12-07 05:41:34.628411] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:01.319 [2024-12-07 05:41:34.628455] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:01.319 [2024-12-07 05:41:34.628465] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:01.319 [2024-12-07 05:41:34.628474] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:01.319 [2024-12-07 05:41:34.628480] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:01.319 [2024-12-07 05:41:34.628488] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.319 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:01.319 "name": "Existed_Raid", 00:14:01.319 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.320 "strip_size_kb": 64, 00:14:01.320 "state": "configuring", 00:14:01.320 "raid_level": "raid5f", 00:14:01.320 "superblock": false, 00:14:01.320 "num_base_bdevs": 4, 00:14:01.320 "num_base_bdevs_discovered": 1, 00:14:01.320 "num_base_bdevs_operational": 4, 00:14:01.320 "base_bdevs_list": [ 00:14:01.320 { 00:14:01.320 "name": "BaseBdev1", 00:14:01.320 "uuid": "ef226877-8793-43a3-a398-13a79349d4d7", 00:14:01.320 "is_configured": true, 00:14:01.320 "data_offset": 0, 00:14:01.320 "data_size": 65536 00:14:01.320 }, 00:14:01.320 { 00:14:01.320 "name": "BaseBdev2", 00:14:01.320 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.320 "is_configured": false, 00:14:01.320 "data_offset": 0, 00:14:01.320 "data_size": 0 00:14:01.320 }, 00:14:01.320 { 00:14:01.320 "name": "BaseBdev3", 00:14:01.320 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.320 "is_configured": false, 00:14:01.320 "data_offset": 0, 00:14:01.320 "data_size": 0 00:14:01.320 }, 00:14:01.320 { 00:14:01.320 "name": "BaseBdev4", 00:14:01.320 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.320 "is_configured": false, 00:14:01.320 "data_offset": 0, 00:14:01.320 "data_size": 0 00:14:01.320 } 00:14:01.320 ] 00:14:01.320 }' 00:14:01.320 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:01.320 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.890 05:41:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:01.890 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.890 05:41:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.890 [2024-12-07 05:41:35.008760] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:01.890 BaseBdev2 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.890 [ 00:14:01.890 { 00:14:01.890 "name": "BaseBdev2", 00:14:01.890 "aliases": [ 00:14:01.890 "61ff8362-abeb-4d7d-baa5-040a9dd7a054" 00:14:01.890 ], 00:14:01.890 "product_name": "Malloc disk", 00:14:01.890 "block_size": 512, 00:14:01.890 "num_blocks": 65536, 00:14:01.890 "uuid": "61ff8362-abeb-4d7d-baa5-040a9dd7a054", 00:14:01.890 "assigned_rate_limits": { 00:14:01.890 "rw_ios_per_sec": 0, 00:14:01.890 "rw_mbytes_per_sec": 0, 00:14:01.890 "r_mbytes_per_sec": 0, 00:14:01.890 "w_mbytes_per_sec": 0 00:14:01.890 }, 00:14:01.890 "claimed": true, 00:14:01.890 "claim_type": "exclusive_write", 00:14:01.890 "zoned": false, 00:14:01.890 "supported_io_types": { 00:14:01.890 "read": true, 00:14:01.890 "write": true, 00:14:01.890 "unmap": true, 00:14:01.890 "flush": true, 00:14:01.890 "reset": true, 00:14:01.890 "nvme_admin": false, 00:14:01.890 "nvme_io": false, 00:14:01.890 "nvme_io_md": false, 00:14:01.890 "write_zeroes": true, 00:14:01.890 "zcopy": true, 00:14:01.890 "get_zone_info": false, 00:14:01.890 "zone_management": false, 00:14:01.890 "zone_append": false, 00:14:01.890 "compare": false, 00:14:01.890 "compare_and_write": false, 00:14:01.890 "abort": true, 00:14:01.890 "seek_hole": false, 00:14:01.890 "seek_data": false, 00:14:01.890 "copy": true, 00:14:01.890 "nvme_iov_md": false 00:14:01.890 }, 00:14:01.890 "memory_domains": [ 00:14:01.890 { 00:14:01.890 "dma_device_id": "system", 00:14:01.890 "dma_device_type": 1 00:14:01.890 }, 00:14:01.890 { 00:14:01.890 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:01.890 "dma_device_type": 2 00:14:01.890 } 00:14:01.890 ], 00:14:01.890 "driver_specific": {} 00:14:01.890 } 00:14:01.890 ] 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:01.890 "name": "Existed_Raid", 00:14:01.890 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.890 "strip_size_kb": 64, 00:14:01.890 "state": "configuring", 00:14:01.890 "raid_level": "raid5f", 00:14:01.890 "superblock": false, 00:14:01.890 "num_base_bdevs": 4, 00:14:01.890 "num_base_bdevs_discovered": 2, 00:14:01.890 "num_base_bdevs_operational": 4, 00:14:01.890 "base_bdevs_list": [ 00:14:01.890 { 00:14:01.890 "name": "BaseBdev1", 00:14:01.890 "uuid": "ef226877-8793-43a3-a398-13a79349d4d7", 00:14:01.890 "is_configured": true, 00:14:01.890 "data_offset": 0, 00:14:01.890 "data_size": 65536 00:14:01.890 }, 00:14:01.890 { 00:14:01.890 "name": "BaseBdev2", 00:14:01.890 "uuid": "61ff8362-abeb-4d7d-baa5-040a9dd7a054", 00:14:01.890 "is_configured": true, 00:14:01.890 "data_offset": 0, 00:14:01.890 "data_size": 65536 00:14:01.890 }, 00:14:01.890 { 00:14:01.890 "name": "BaseBdev3", 00:14:01.890 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.890 "is_configured": false, 00:14:01.890 "data_offset": 0, 00:14:01.890 "data_size": 0 00:14:01.890 }, 00:14:01.890 { 00:14:01.890 "name": "BaseBdev4", 00:14:01.890 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.890 "is_configured": false, 00:14:01.890 "data_offset": 0, 00:14:01.890 "data_size": 0 00:14:01.890 } 00:14:01.890 ] 00:14:01.890 }' 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:01.890 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.149 [2024-12-07 05:41:35.446469] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:02.149 BaseBdev3 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.149 [ 00:14:02.149 { 00:14:02.149 "name": "BaseBdev3", 00:14:02.149 "aliases": [ 00:14:02.149 "6827bf97-87ab-4d8f-95e5-a9025bfa4774" 00:14:02.149 ], 00:14:02.149 "product_name": "Malloc disk", 00:14:02.149 "block_size": 512, 00:14:02.149 "num_blocks": 65536, 00:14:02.149 "uuid": "6827bf97-87ab-4d8f-95e5-a9025bfa4774", 00:14:02.149 "assigned_rate_limits": { 00:14:02.149 "rw_ios_per_sec": 0, 00:14:02.149 "rw_mbytes_per_sec": 0, 00:14:02.149 "r_mbytes_per_sec": 0, 00:14:02.149 "w_mbytes_per_sec": 0 00:14:02.149 }, 00:14:02.149 "claimed": true, 00:14:02.149 "claim_type": "exclusive_write", 00:14:02.149 "zoned": false, 00:14:02.149 "supported_io_types": { 00:14:02.149 "read": true, 00:14:02.149 "write": true, 00:14:02.149 "unmap": true, 00:14:02.149 "flush": true, 00:14:02.149 "reset": true, 00:14:02.149 "nvme_admin": false, 00:14:02.149 "nvme_io": false, 00:14:02.149 "nvme_io_md": false, 00:14:02.149 "write_zeroes": true, 00:14:02.149 "zcopy": true, 00:14:02.149 "get_zone_info": false, 00:14:02.149 "zone_management": false, 00:14:02.149 "zone_append": false, 00:14:02.149 "compare": false, 00:14:02.149 "compare_and_write": false, 00:14:02.149 "abort": true, 00:14:02.149 "seek_hole": false, 00:14:02.149 "seek_data": false, 00:14:02.149 "copy": true, 00:14:02.149 "nvme_iov_md": false 00:14:02.149 }, 00:14:02.149 "memory_domains": [ 00:14:02.149 { 00:14:02.149 "dma_device_id": "system", 00:14:02.149 "dma_device_type": 1 00:14:02.149 }, 00:14:02.149 { 00:14:02.149 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:02.149 "dma_device_type": 2 00:14:02.149 } 00:14:02.149 ], 00:14:02.149 "driver_specific": {} 00:14:02.149 } 00:14:02.149 ] 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:02.149 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:02.150 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:02.150 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:02.150 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:02.150 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.150 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.150 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.150 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:02.150 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.409 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:02.409 "name": "Existed_Raid", 00:14:02.409 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:02.409 "strip_size_kb": 64, 00:14:02.409 "state": "configuring", 00:14:02.409 "raid_level": "raid5f", 00:14:02.409 "superblock": false, 00:14:02.409 "num_base_bdevs": 4, 00:14:02.409 "num_base_bdevs_discovered": 3, 00:14:02.409 "num_base_bdevs_operational": 4, 00:14:02.409 "base_bdevs_list": [ 00:14:02.409 { 00:14:02.409 "name": "BaseBdev1", 00:14:02.409 "uuid": "ef226877-8793-43a3-a398-13a79349d4d7", 00:14:02.409 "is_configured": true, 00:14:02.409 "data_offset": 0, 00:14:02.409 "data_size": 65536 00:14:02.409 }, 00:14:02.409 { 00:14:02.409 "name": "BaseBdev2", 00:14:02.409 "uuid": "61ff8362-abeb-4d7d-baa5-040a9dd7a054", 00:14:02.409 "is_configured": true, 00:14:02.409 "data_offset": 0, 00:14:02.409 "data_size": 65536 00:14:02.409 }, 00:14:02.409 { 00:14:02.409 "name": "BaseBdev3", 00:14:02.409 "uuid": "6827bf97-87ab-4d8f-95e5-a9025bfa4774", 00:14:02.409 "is_configured": true, 00:14:02.409 "data_offset": 0, 00:14:02.409 "data_size": 65536 00:14:02.409 }, 00:14:02.409 { 00:14:02.409 "name": "BaseBdev4", 00:14:02.409 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:02.409 "is_configured": false, 00:14:02.409 "data_offset": 0, 00:14:02.409 "data_size": 0 00:14:02.409 } 00:14:02.409 ] 00:14:02.409 }' 00:14:02.409 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:02.409 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.668 [2024-12-07 05:41:35.916656] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:02.668 [2024-12-07 05:41:35.916795] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:02.668 [2024-12-07 05:41:35.916831] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:02.668 [2024-12-07 05:41:35.917138] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:02.668 [2024-12-07 05:41:35.917727] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:02.668 [2024-12-07 05:41:35.917783] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:14:02.668 [2024-12-07 05:41:35.918018] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:02.668 BaseBdev4 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.668 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.669 [ 00:14:02.669 { 00:14:02.669 "name": "BaseBdev4", 00:14:02.669 "aliases": [ 00:14:02.669 "99e22ab7-dc4d-4a2b-9b58-a765a3aef4ee" 00:14:02.669 ], 00:14:02.669 "product_name": "Malloc disk", 00:14:02.669 "block_size": 512, 00:14:02.669 "num_blocks": 65536, 00:14:02.669 "uuid": "99e22ab7-dc4d-4a2b-9b58-a765a3aef4ee", 00:14:02.669 "assigned_rate_limits": { 00:14:02.669 "rw_ios_per_sec": 0, 00:14:02.669 "rw_mbytes_per_sec": 0, 00:14:02.669 "r_mbytes_per_sec": 0, 00:14:02.669 "w_mbytes_per_sec": 0 00:14:02.669 }, 00:14:02.669 "claimed": true, 00:14:02.669 "claim_type": "exclusive_write", 00:14:02.669 "zoned": false, 00:14:02.669 "supported_io_types": { 00:14:02.669 "read": true, 00:14:02.669 "write": true, 00:14:02.669 "unmap": true, 00:14:02.669 "flush": true, 00:14:02.669 "reset": true, 00:14:02.669 "nvme_admin": false, 00:14:02.669 "nvme_io": false, 00:14:02.669 "nvme_io_md": false, 00:14:02.669 "write_zeroes": true, 00:14:02.669 "zcopy": true, 00:14:02.669 "get_zone_info": false, 00:14:02.669 "zone_management": false, 00:14:02.669 "zone_append": false, 00:14:02.669 "compare": false, 00:14:02.669 "compare_and_write": false, 00:14:02.669 "abort": true, 00:14:02.669 "seek_hole": false, 00:14:02.669 "seek_data": false, 00:14:02.669 "copy": true, 00:14:02.669 "nvme_iov_md": false 00:14:02.669 }, 00:14:02.669 "memory_domains": [ 00:14:02.669 { 00:14:02.669 "dma_device_id": "system", 00:14:02.669 "dma_device_type": 1 00:14:02.669 }, 00:14:02.669 { 00:14:02.669 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:02.669 "dma_device_type": 2 00:14:02.669 } 00:14:02.669 ], 00:14:02.669 "driver_specific": {} 00:14:02.669 } 00:14:02.669 ] 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:02.669 "name": "Existed_Raid", 00:14:02.669 "uuid": "b366d540-5f92-4dad-8877-8d9cf0410432", 00:14:02.669 "strip_size_kb": 64, 00:14:02.669 "state": "online", 00:14:02.669 "raid_level": "raid5f", 00:14:02.669 "superblock": false, 00:14:02.669 "num_base_bdevs": 4, 00:14:02.669 "num_base_bdevs_discovered": 4, 00:14:02.669 "num_base_bdevs_operational": 4, 00:14:02.669 "base_bdevs_list": [ 00:14:02.669 { 00:14:02.669 "name": "BaseBdev1", 00:14:02.669 "uuid": "ef226877-8793-43a3-a398-13a79349d4d7", 00:14:02.669 "is_configured": true, 00:14:02.669 "data_offset": 0, 00:14:02.669 "data_size": 65536 00:14:02.669 }, 00:14:02.669 { 00:14:02.669 "name": "BaseBdev2", 00:14:02.669 "uuid": "61ff8362-abeb-4d7d-baa5-040a9dd7a054", 00:14:02.669 "is_configured": true, 00:14:02.669 "data_offset": 0, 00:14:02.669 "data_size": 65536 00:14:02.669 }, 00:14:02.669 { 00:14:02.669 "name": "BaseBdev3", 00:14:02.669 "uuid": "6827bf97-87ab-4d8f-95e5-a9025bfa4774", 00:14:02.669 "is_configured": true, 00:14:02.669 "data_offset": 0, 00:14:02.669 "data_size": 65536 00:14:02.669 }, 00:14:02.669 { 00:14:02.669 "name": "BaseBdev4", 00:14:02.669 "uuid": "99e22ab7-dc4d-4a2b-9b58-a765a3aef4ee", 00:14:02.669 "is_configured": true, 00:14:02.669 "data_offset": 0, 00:14:02.669 "data_size": 65536 00:14:02.669 } 00:14:02.669 ] 00:14:02.669 }' 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:02.669 05:41:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:03.240 [2024-12-07 05:41:36.412006] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:03.240 "name": "Existed_Raid", 00:14:03.240 "aliases": [ 00:14:03.240 "b366d540-5f92-4dad-8877-8d9cf0410432" 00:14:03.240 ], 00:14:03.240 "product_name": "Raid Volume", 00:14:03.240 "block_size": 512, 00:14:03.240 "num_blocks": 196608, 00:14:03.240 "uuid": "b366d540-5f92-4dad-8877-8d9cf0410432", 00:14:03.240 "assigned_rate_limits": { 00:14:03.240 "rw_ios_per_sec": 0, 00:14:03.240 "rw_mbytes_per_sec": 0, 00:14:03.240 "r_mbytes_per_sec": 0, 00:14:03.240 "w_mbytes_per_sec": 0 00:14:03.240 }, 00:14:03.240 "claimed": false, 00:14:03.240 "zoned": false, 00:14:03.240 "supported_io_types": { 00:14:03.240 "read": true, 00:14:03.240 "write": true, 00:14:03.240 "unmap": false, 00:14:03.240 "flush": false, 00:14:03.240 "reset": true, 00:14:03.240 "nvme_admin": false, 00:14:03.240 "nvme_io": false, 00:14:03.240 "nvme_io_md": false, 00:14:03.240 "write_zeroes": true, 00:14:03.240 "zcopy": false, 00:14:03.240 "get_zone_info": false, 00:14:03.240 "zone_management": false, 00:14:03.240 "zone_append": false, 00:14:03.240 "compare": false, 00:14:03.240 "compare_and_write": false, 00:14:03.240 "abort": false, 00:14:03.240 "seek_hole": false, 00:14:03.240 "seek_data": false, 00:14:03.240 "copy": false, 00:14:03.240 "nvme_iov_md": false 00:14:03.240 }, 00:14:03.240 "driver_specific": { 00:14:03.240 "raid": { 00:14:03.240 "uuid": "b366d540-5f92-4dad-8877-8d9cf0410432", 00:14:03.240 "strip_size_kb": 64, 00:14:03.240 "state": "online", 00:14:03.240 "raid_level": "raid5f", 00:14:03.240 "superblock": false, 00:14:03.240 "num_base_bdevs": 4, 00:14:03.240 "num_base_bdevs_discovered": 4, 00:14:03.240 "num_base_bdevs_operational": 4, 00:14:03.240 "base_bdevs_list": [ 00:14:03.240 { 00:14:03.240 "name": "BaseBdev1", 00:14:03.240 "uuid": "ef226877-8793-43a3-a398-13a79349d4d7", 00:14:03.240 "is_configured": true, 00:14:03.240 "data_offset": 0, 00:14:03.240 "data_size": 65536 00:14:03.240 }, 00:14:03.240 { 00:14:03.240 "name": "BaseBdev2", 00:14:03.240 "uuid": "61ff8362-abeb-4d7d-baa5-040a9dd7a054", 00:14:03.240 "is_configured": true, 00:14:03.240 "data_offset": 0, 00:14:03.240 "data_size": 65536 00:14:03.240 }, 00:14:03.240 { 00:14:03.240 "name": "BaseBdev3", 00:14:03.240 "uuid": "6827bf97-87ab-4d8f-95e5-a9025bfa4774", 00:14:03.240 "is_configured": true, 00:14:03.240 "data_offset": 0, 00:14:03.240 "data_size": 65536 00:14:03.240 }, 00:14:03.240 { 00:14:03.240 "name": "BaseBdev4", 00:14:03.240 "uuid": "99e22ab7-dc4d-4a2b-9b58-a765a3aef4ee", 00:14:03.240 "is_configured": true, 00:14:03.240 "data_offset": 0, 00:14:03.240 "data_size": 65536 00:14:03.240 } 00:14:03.240 ] 00:14:03.240 } 00:14:03.240 } 00:14:03.240 }' 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:03.240 BaseBdev2 00:14:03.240 BaseBdev3 00:14:03.240 BaseBdev4' 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:03.240 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:03.501 [2024-12-07 05:41:36.703386] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.501 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:03.501 "name": "Existed_Raid", 00:14:03.501 "uuid": "b366d540-5f92-4dad-8877-8d9cf0410432", 00:14:03.501 "strip_size_kb": 64, 00:14:03.501 "state": "online", 00:14:03.501 "raid_level": "raid5f", 00:14:03.501 "superblock": false, 00:14:03.501 "num_base_bdevs": 4, 00:14:03.501 "num_base_bdevs_discovered": 3, 00:14:03.501 "num_base_bdevs_operational": 3, 00:14:03.501 "base_bdevs_list": [ 00:14:03.501 { 00:14:03.501 "name": null, 00:14:03.501 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:03.501 "is_configured": false, 00:14:03.501 "data_offset": 0, 00:14:03.501 "data_size": 65536 00:14:03.501 }, 00:14:03.501 { 00:14:03.501 "name": "BaseBdev2", 00:14:03.501 "uuid": "61ff8362-abeb-4d7d-baa5-040a9dd7a054", 00:14:03.501 "is_configured": true, 00:14:03.502 "data_offset": 0, 00:14:03.502 "data_size": 65536 00:14:03.502 }, 00:14:03.502 { 00:14:03.502 "name": "BaseBdev3", 00:14:03.502 "uuid": "6827bf97-87ab-4d8f-95e5-a9025bfa4774", 00:14:03.502 "is_configured": true, 00:14:03.502 "data_offset": 0, 00:14:03.502 "data_size": 65536 00:14:03.502 }, 00:14:03.502 { 00:14:03.502 "name": "BaseBdev4", 00:14:03.502 "uuid": "99e22ab7-dc4d-4a2b-9b58-a765a3aef4ee", 00:14:03.502 "is_configured": true, 00:14:03.502 "data_offset": 0, 00:14:03.502 "data_size": 65536 00:14:03.502 } 00:14:03.502 ] 00:14:03.502 }' 00:14:03.502 05:41:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:03.502 05:41:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.072 [2024-12-07 05:41:37.193515] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:04.072 [2024-12-07 05:41:37.193613] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:04.072 [2024-12-07 05:41:37.204667] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.072 [2024-12-07 05:41:37.264591] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.072 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.073 [2024-12-07 05:41:37.327634] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:04.073 [2024-12-07 05:41:37.327674] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.073 BaseBdev2 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.073 [ 00:14:04.073 { 00:14:04.073 "name": "BaseBdev2", 00:14:04.073 "aliases": [ 00:14:04.073 "ddd7f038-46e7-46ef-b5ec-388222005ef2" 00:14:04.073 ], 00:14:04.073 "product_name": "Malloc disk", 00:14:04.073 "block_size": 512, 00:14:04.073 "num_blocks": 65536, 00:14:04.073 "uuid": "ddd7f038-46e7-46ef-b5ec-388222005ef2", 00:14:04.073 "assigned_rate_limits": { 00:14:04.073 "rw_ios_per_sec": 0, 00:14:04.073 "rw_mbytes_per_sec": 0, 00:14:04.073 "r_mbytes_per_sec": 0, 00:14:04.073 "w_mbytes_per_sec": 0 00:14:04.073 }, 00:14:04.073 "claimed": false, 00:14:04.073 "zoned": false, 00:14:04.073 "supported_io_types": { 00:14:04.073 "read": true, 00:14:04.073 "write": true, 00:14:04.073 "unmap": true, 00:14:04.073 "flush": true, 00:14:04.073 "reset": true, 00:14:04.073 "nvme_admin": false, 00:14:04.073 "nvme_io": false, 00:14:04.073 "nvme_io_md": false, 00:14:04.073 "write_zeroes": true, 00:14:04.073 "zcopy": true, 00:14:04.073 "get_zone_info": false, 00:14:04.073 "zone_management": false, 00:14:04.073 "zone_append": false, 00:14:04.073 "compare": false, 00:14:04.073 "compare_and_write": false, 00:14:04.073 "abort": true, 00:14:04.073 "seek_hole": false, 00:14:04.073 "seek_data": false, 00:14:04.073 "copy": true, 00:14:04.073 "nvme_iov_md": false 00:14:04.073 }, 00:14:04.073 "memory_domains": [ 00:14:04.073 { 00:14:04.073 "dma_device_id": "system", 00:14:04.073 "dma_device_type": 1 00:14:04.073 }, 00:14:04.073 { 00:14:04.073 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:04.073 "dma_device_type": 2 00:14:04.073 } 00:14:04.073 ], 00:14:04.073 "driver_specific": {} 00:14:04.073 } 00:14:04.073 ] 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.073 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.333 BaseBdev3 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.333 [ 00:14:04.333 { 00:14:04.333 "name": "BaseBdev3", 00:14:04.333 "aliases": [ 00:14:04.333 "b74ee733-3fdf-447b-a800-76acb45f8c66" 00:14:04.333 ], 00:14:04.333 "product_name": "Malloc disk", 00:14:04.333 "block_size": 512, 00:14:04.333 "num_blocks": 65536, 00:14:04.333 "uuid": "b74ee733-3fdf-447b-a800-76acb45f8c66", 00:14:04.333 "assigned_rate_limits": { 00:14:04.333 "rw_ios_per_sec": 0, 00:14:04.333 "rw_mbytes_per_sec": 0, 00:14:04.333 "r_mbytes_per_sec": 0, 00:14:04.333 "w_mbytes_per_sec": 0 00:14:04.333 }, 00:14:04.333 "claimed": false, 00:14:04.333 "zoned": false, 00:14:04.333 "supported_io_types": { 00:14:04.333 "read": true, 00:14:04.333 "write": true, 00:14:04.333 "unmap": true, 00:14:04.333 "flush": true, 00:14:04.333 "reset": true, 00:14:04.333 "nvme_admin": false, 00:14:04.333 "nvme_io": false, 00:14:04.333 "nvme_io_md": false, 00:14:04.333 "write_zeroes": true, 00:14:04.333 "zcopy": true, 00:14:04.333 "get_zone_info": false, 00:14:04.333 "zone_management": false, 00:14:04.333 "zone_append": false, 00:14:04.333 "compare": false, 00:14:04.333 "compare_and_write": false, 00:14:04.333 "abort": true, 00:14:04.333 "seek_hole": false, 00:14:04.333 "seek_data": false, 00:14:04.333 "copy": true, 00:14:04.333 "nvme_iov_md": false 00:14:04.333 }, 00:14:04.333 "memory_domains": [ 00:14:04.333 { 00:14:04.333 "dma_device_id": "system", 00:14:04.333 "dma_device_type": 1 00:14:04.333 }, 00:14:04.333 { 00:14:04.333 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:04.333 "dma_device_type": 2 00:14:04.333 } 00:14:04.333 ], 00:14:04.333 "driver_specific": {} 00:14:04.333 } 00:14:04.333 ] 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.333 BaseBdev4 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.333 [ 00:14:04.333 { 00:14:04.333 "name": "BaseBdev4", 00:14:04.333 "aliases": [ 00:14:04.333 "a844086a-c9ce-4f4e-aa85-d68412c7ebd3" 00:14:04.333 ], 00:14:04.333 "product_name": "Malloc disk", 00:14:04.333 "block_size": 512, 00:14:04.333 "num_blocks": 65536, 00:14:04.333 "uuid": "a844086a-c9ce-4f4e-aa85-d68412c7ebd3", 00:14:04.333 "assigned_rate_limits": { 00:14:04.333 "rw_ios_per_sec": 0, 00:14:04.333 "rw_mbytes_per_sec": 0, 00:14:04.333 "r_mbytes_per_sec": 0, 00:14:04.333 "w_mbytes_per_sec": 0 00:14:04.333 }, 00:14:04.333 "claimed": false, 00:14:04.333 "zoned": false, 00:14:04.333 "supported_io_types": { 00:14:04.333 "read": true, 00:14:04.333 "write": true, 00:14:04.333 "unmap": true, 00:14:04.333 "flush": true, 00:14:04.333 "reset": true, 00:14:04.333 "nvme_admin": false, 00:14:04.333 "nvme_io": false, 00:14:04.333 "nvme_io_md": false, 00:14:04.333 "write_zeroes": true, 00:14:04.333 "zcopy": true, 00:14:04.333 "get_zone_info": false, 00:14:04.333 "zone_management": false, 00:14:04.333 "zone_append": false, 00:14:04.333 "compare": false, 00:14:04.333 "compare_and_write": false, 00:14:04.333 "abort": true, 00:14:04.333 "seek_hole": false, 00:14:04.333 "seek_data": false, 00:14:04.333 "copy": true, 00:14:04.333 "nvme_iov_md": false 00:14:04.333 }, 00:14:04.333 "memory_domains": [ 00:14:04.333 { 00:14:04.333 "dma_device_id": "system", 00:14:04.333 "dma_device_type": 1 00:14:04.333 }, 00:14:04.333 { 00:14:04.333 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:04.333 "dma_device_type": 2 00:14:04.333 } 00:14:04.333 ], 00:14:04.333 "driver_specific": {} 00:14:04.333 } 00:14:04.333 ] 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.333 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.333 [2024-12-07 05:41:37.542076] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:04.333 [2024-12-07 05:41:37.542143] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:04.334 [2024-12-07 05:41:37.542169] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:04.334 [2024-12-07 05:41:37.543946] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:04.334 [2024-12-07 05:41:37.544012] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:04.334 "name": "Existed_Raid", 00:14:04.334 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:04.334 "strip_size_kb": 64, 00:14:04.334 "state": "configuring", 00:14:04.334 "raid_level": "raid5f", 00:14:04.334 "superblock": false, 00:14:04.334 "num_base_bdevs": 4, 00:14:04.334 "num_base_bdevs_discovered": 3, 00:14:04.334 "num_base_bdevs_operational": 4, 00:14:04.334 "base_bdevs_list": [ 00:14:04.334 { 00:14:04.334 "name": "BaseBdev1", 00:14:04.334 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:04.334 "is_configured": false, 00:14:04.334 "data_offset": 0, 00:14:04.334 "data_size": 0 00:14:04.334 }, 00:14:04.334 { 00:14:04.334 "name": "BaseBdev2", 00:14:04.334 "uuid": "ddd7f038-46e7-46ef-b5ec-388222005ef2", 00:14:04.334 "is_configured": true, 00:14:04.334 "data_offset": 0, 00:14:04.334 "data_size": 65536 00:14:04.334 }, 00:14:04.334 { 00:14:04.334 "name": "BaseBdev3", 00:14:04.334 "uuid": "b74ee733-3fdf-447b-a800-76acb45f8c66", 00:14:04.334 "is_configured": true, 00:14:04.334 "data_offset": 0, 00:14:04.334 "data_size": 65536 00:14:04.334 }, 00:14:04.334 { 00:14:04.334 "name": "BaseBdev4", 00:14:04.334 "uuid": "a844086a-c9ce-4f4e-aa85-d68412c7ebd3", 00:14:04.334 "is_configured": true, 00:14:04.334 "data_offset": 0, 00:14:04.334 "data_size": 65536 00:14:04.334 } 00:14:04.334 ] 00:14:04.334 }' 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:04.334 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.903 [2024-12-07 05:41:37.985278] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.903 05:41:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:04.903 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.903 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:04.903 "name": "Existed_Raid", 00:14:04.903 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:04.903 "strip_size_kb": 64, 00:14:04.903 "state": "configuring", 00:14:04.903 "raid_level": "raid5f", 00:14:04.903 "superblock": false, 00:14:04.903 "num_base_bdevs": 4, 00:14:04.903 "num_base_bdevs_discovered": 2, 00:14:04.903 "num_base_bdevs_operational": 4, 00:14:04.903 "base_bdevs_list": [ 00:14:04.903 { 00:14:04.903 "name": "BaseBdev1", 00:14:04.903 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:04.903 "is_configured": false, 00:14:04.904 "data_offset": 0, 00:14:04.904 "data_size": 0 00:14:04.904 }, 00:14:04.904 { 00:14:04.904 "name": null, 00:14:04.904 "uuid": "ddd7f038-46e7-46ef-b5ec-388222005ef2", 00:14:04.904 "is_configured": false, 00:14:04.904 "data_offset": 0, 00:14:04.904 "data_size": 65536 00:14:04.904 }, 00:14:04.904 { 00:14:04.904 "name": "BaseBdev3", 00:14:04.904 "uuid": "b74ee733-3fdf-447b-a800-76acb45f8c66", 00:14:04.904 "is_configured": true, 00:14:04.904 "data_offset": 0, 00:14:04.904 "data_size": 65536 00:14:04.904 }, 00:14:04.904 { 00:14:04.904 "name": "BaseBdev4", 00:14:04.904 "uuid": "a844086a-c9ce-4f4e-aa85-d68412c7ebd3", 00:14:04.904 "is_configured": true, 00:14:04.904 "data_offset": 0, 00:14:04.904 "data_size": 65536 00:14:04.904 } 00:14:04.904 ] 00:14:04.904 }' 00:14:04.904 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:04.904 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.164 [2024-12-07 05:41:38.491361] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:05.164 BaseBdev1 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.164 [ 00:14:05.164 { 00:14:05.164 "name": "BaseBdev1", 00:14:05.164 "aliases": [ 00:14:05.164 "36601767-4bb2-4ddb-a0fa-de40780f6ce4" 00:14:05.164 ], 00:14:05.164 "product_name": "Malloc disk", 00:14:05.164 "block_size": 512, 00:14:05.164 "num_blocks": 65536, 00:14:05.164 "uuid": "36601767-4bb2-4ddb-a0fa-de40780f6ce4", 00:14:05.164 "assigned_rate_limits": { 00:14:05.164 "rw_ios_per_sec": 0, 00:14:05.164 "rw_mbytes_per_sec": 0, 00:14:05.164 "r_mbytes_per_sec": 0, 00:14:05.164 "w_mbytes_per_sec": 0 00:14:05.164 }, 00:14:05.164 "claimed": true, 00:14:05.164 "claim_type": "exclusive_write", 00:14:05.164 "zoned": false, 00:14:05.164 "supported_io_types": { 00:14:05.164 "read": true, 00:14:05.164 "write": true, 00:14:05.164 "unmap": true, 00:14:05.164 "flush": true, 00:14:05.164 "reset": true, 00:14:05.164 "nvme_admin": false, 00:14:05.164 "nvme_io": false, 00:14:05.164 "nvme_io_md": false, 00:14:05.164 "write_zeroes": true, 00:14:05.164 "zcopy": true, 00:14:05.164 "get_zone_info": false, 00:14:05.164 "zone_management": false, 00:14:05.164 "zone_append": false, 00:14:05.164 "compare": false, 00:14:05.164 "compare_and_write": false, 00:14:05.164 "abort": true, 00:14:05.164 "seek_hole": false, 00:14:05.164 "seek_data": false, 00:14:05.164 "copy": true, 00:14:05.164 "nvme_iov_md": false 00:14:05.164 }, 00:14:05.164 "memory_domains": [ 00:14:05.164 { 00:14:05.164 "dma_device_id": "system", 00:14:05.164 "dma_device_type": 1 00:14:05.164 }, 00:14:05.164 { 00:14:05.164 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:05.164 "dma_device_type": 2 00:14:05.164 } 00:14:05.164 ], 00:14:05.164 "driver_specific": {} 00:14:05.164 } 00:14:05.164 ] 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:05.164 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:05.424 "name": "Existed_Raid", 00:14:05.424 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:05.424 "strip_size_kb": 64, 00:14:05.424 "state": "configuring", 00:14:05.424 "raid_level": "raid5f", 00:14:05.424 "superblock": false, 00:14:05.424 "num_base_bdevs": 4, 00:14:05.424 "num_base_bdevs_discovered": 3, 00:14:05.424 "num_base_bdevs_operational": 4, 00:14:05.424 "base_bdevs_list": [ 00:14:05.424 { 00:14:05.424 "name": "BaseBdev1", 00:14:05.424 "uuid": "36601767-4bb2-4ddb-a0fa-de40780f6ce4", 00:14:05.424 "is_configured": true, 00:14:05.424 "data_offset": 0, 00:14:05.424 "data_size": 65536 00:14:05.424 }, 00:14:05.424 { 00:14:05.424 "name": null, 00:14:05.424 "uuid": "ddd7f038-46e7-46ef-b5ec-388222005ef2", 00:14:05.424 "is_configured": false, 00:14:05.424 "data_offset": 0, 00:14:05.424 "data_size": 65536 00:14:05.424 }, 00:14:05.424 { 00:14:05.424 "name": "BaseBdev3", 00:14:05.424 "uuid": "b74ee733-3fdf-447b-a800-76acb45f8c66", 00:14:05.424 "is_configured": true, 00:14:05.424 "data_offset": 0, 00:14:05.424 "data_size": 65536 00:14:05.424 }, 00:14:05.424 { 00:14:05.424 "name": "BaseBdev4", 00:14:05.424 "uuid": "a844086a-c9ce-4f4e-aa85-d68412c7ebd3", 00:14:05.424 "is_configured": true, 00:14:05.424 "data_offset": 0, 00:14:05.424 "data_size": 65536 00:14:05.424 } 00:14:05.424 ] 00:14:05.424 }' 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:05.424 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.684 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.684 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.684 05:41:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:05.684 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.684 05:41:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.684 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:05.684 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.685 [2024-12-07 05:41:39.026508] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.685 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:05.944 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.944 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:05.944 "name": "Existed_Raid", 00:14:05.944 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:05.944 "strip_size_kb": 64, 00:14:05.944 "state": "configuring", 00:14:05.944 "raid_level": "raid5f", 00:14:05.944 "superblock": false, 00:14:05.944 "num_base_bdevs": 4, 00:14:05.944 "num_base_bdevs_discovered": 2, 00:14:05.944 "num_base_bdevs_operational": 4, 00:14:05.944 "base_bdevs_list": [ 00:14:05.944 { 00:14:05.944 "name": "BaseBdev1", 00:14:05.944 "uuid": "36601767-4bb2-4ddb-a0fa-de40780f6ce4", 00:14:05.944 "is_configured": true, 00:14:05.944 "data_offset": 0, 00:14:05.944 "data_size": 65536 00:14:05.944 }, 00:14:05.945 { 00:14:05.945 "name": null, 00:14:05.945 "uuid": "ddd7f038-46e7-46ef-b5ec-388222005ef2", 00:14:05.945 "is_configured": false, 00:14:05.945 "data_offset": 0, 00:14:05.945 "data_size": 65536 00:14:05.945 }, 00:14:05.945 { 00:14:05.945 "name": null, 00:14:05.945 "uuid": "b74ee733-3fdf-447b-a800-76acb45f8c66", 00:14:05.945 "is_configured": false, 00:14:05.945 "data_offset": 0, 00:14:05.945 "data_size": 65536 00:14:05.945 }, 00:14:05.945 { 00:14:05.945 "name": "BaseBdev4", 00:14:05.945 "uuid": "a844086a-c9ce-4f4e-aa85-d68412c7ebd3", 00:14:05.945 "is_configured": true, 00:14:05.945 "data_offset": 0, 00:14:05.945 "data_size": 65536 00:14:05.945 } 00:14:05.945 ] 00:14:05.945 }' 00:14:05.945 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:05.945 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.205 [2024-12-07 05:41:39.457791] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:06.205 "name": "Existed_Raid", 00:14:06.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:06.205 "strip_size_kb": 64, 00:14:06.205 "state": "configuring", 00:14:06.205 "raid_level": "raid5f", 00:14:06.205 "superblock": false, 00:14:06.205 "num_base_bdevs": 4, 00:14:06.205 "num_base_bdevs_discovered": 3, 00:14:06.205 "num_base_bdevs_operational": 4, 00:14:06.205 "base_bdevs_list": [ 00:14:06.205 { 00:14:06.205 "name": "BaseBdev1", 00:14:06.205 "uuid": "36601767-4bb2-4ddb-a0fa-de40780f6ce4", 00:14:06.205 "is_configured": true, 00:14:06.205 "data_offset": 0, 00:14:06.205 "data_size": 65536 00:14:06.205 }, 00:14:06.205 { 00:14:06.205 "name": null, 00:14:06.205 "uuid": "ddd7f038-46e7-46ef-b5ec-388222005ef2", 00:14:06.205 "is_configured": false, 00:14:06.205 "data_offset": 0, 00:14:06.205 "data_size": 65536 00:14:06.205 }, 00:14:06.205 { 00:14:06.205 "name": "BaseBdev3", 00:14:06.205 "uuid": "b74ee733-3fdf-447b-a800-76acb45f8c66", 00:14:06.205 "is_configured": true, 00:14:06.205 "data_offset": 0, 00:14:06.205 "data_size": 65536 00:14:06.205 }, 00:14:06.205 { 00:14:06.205 "name": "BaseBdev4", 00:14:06.205 "uuid": "a844086a-c9ce-4f4e-aa85-d68412c7ebd3", 00:14:06.205 "is_configured": true, 00:14:06.205 "data_offset": 0, 00:14:06.205 "data_size": 65536 00:14:06.205 } 00:14:06.205 ] 00:14:06.205 }' 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:06.205 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.773 [2024-12-07 05:41:39.881070] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.773 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:06.773 "name": "Existed_Raid", 00:14:06.773 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:06.773 "strip_size_kb": 64, 00:14:06.773 "state": "configuring", 00:14:06.773 "raid_level": "raid5f", 00:14:06.773 "superblock": false, 00:14:06.773 "num_base_bdevs": 4, 00:14:06.773 "num_base_bdevs_discovered": 2, 00:14:06.773 "num_base_bdevs_operational": 4, 00:14:06.773 "base_bdevs_list": [ 00:14:06.773 { 00:14:06.773 "name": null, 00:14:06.773 "uuid": "36601767-4bb2-4ddb-a0fa-de40780f6ce4", 00:14:06.773 "is_configured": false, 00:14:06.773 "data_offset": 0, 00:14:06.773 "data_size": 65536 00:14:06.773 }, 00:14:06.773 { 00:14:06.773 "name": null, 00:14:06.773 "uuid": "ddd7f038-46e7-46ef-b5ec-388222005ef2", 00:14:06.773 "is_configured": false, 00:14:06.773 "data_offset": 0, 00:14:06.773 "data_size": 65536 00:14:06.773 }, 00:14:06.773 { 00:14:06.773 "name": "BaseBdev3", 00:14:06.773 "uuid": "b74ee733-3fdf-447b-a800-76acb45f8c66", 00:14:06.773 "is_configured": true, 00:14:06.774 "data_offset": 0, 00:14:06.774 "data_size": 65536 00:14:06.774 }, 00:14:06.774 { 00:14:06.774 "name": "BaseBdev4", 00:14:06.774 "uuid": "a844086a-c9ce-4f4e-aa85-d68412c7ebd3", 00:14:06.774 "is_configured": true, 00:14:06.774 "data_offset": 0, 00:14:06.774 "data_size": 65536 00:14:06.774 } 00:14:06.774 ] 00:14:06.774 }' 00:14:06.774 05:41:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:06.774 05:41:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.033 [2024-12-07 05:41:40.350493] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.033 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.292 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:07.292 "name": "Existed_Raid", 00:14:07.292 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:07.292 "strip_size_kb": 64, 00:14:07.292 "state": "configuring", 00:14:07.292 "raid_level": "raid5f", 00:14:07.292 "superblock": false, 00:14:07.292 "num_base_bdevs": 4, 00:14:07.292 "num_base_bdevs_discovered": 3, 00:14:07.292 "num_base_bdevs_operational": 4, 00:14:07.292 "base_bdevs_list": [ 00:14:07.292 { 00:14:07.292 "name": null, 00:14:07.292 "uuid": "36601767-4bb2-4ddb-a0fa-de40780f6ce4", 00:14:07.292 "is_configured": false, 00:14:07.292 "data_offset": 0, 00:14:07.292 "data_size": 65536 00:14:07.292 }, 00:14:07.292 { 00:14:07.292 "name": "BaseBdev2", 00:14:07.292 "uuid": "ddd7f038-46e7-46ef-b5ec-388222005ef2", 00:14:07.292 "is_configured": true, 00:14:07.292 "data_offset": 0, 00:14:07.292 "data_size": 65536 00:14:07.292 }, 00:14:07.292 { 00:14:07.292 "name": "BaseBdev3", 00:14:07.292 "uuid": "b74ee733-3fdf-447b-a800-76acb45f8c66", 00:14:07.292 "is_configured": true, 00:14:07.292 "data_offset": 0, 00:14:07.292 "data_size": 65536 00:14:07.292 }, 00:14:07.292 { 00:14:07.292 "name": "BaseBdev4", 00:14:07.292 "uuid": "a844086a-c9ce-4f4e-aa85-d68412c7ebd3", 00:14:07.292 "is_configured": true, 00:14:07.292 "data_offset": 0, 00:14:07.292 "data_size": 65536 00:14:07.292 } 00:14:07.292 ] 00:14:07.292 }' 00:14:07.292 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:07.292 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.551 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 36601767-4bb2-4ddb-a0fa-de40780f6ce4 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.552 [2024-12-07 05:41:40.880401] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:07.552 [2024-12-07 05:41:40.880453] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:07.552 [2024-12-07 05:41:40.880462] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:07.552 [2024-12-07 05:41:40.880733] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:14:07.552 [2024-12-07 05:41:40.881179] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:07.552 [2024-12-07 05:41:40.881201] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:14:07.552 [2024-12-07 05:41:40.881375] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:07.552 NewBaseBdev 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.552 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.552 [ 00:14:07.552 { 00:14:07.552 "name": "NewBaseBdev", 00:14:07.552 "aliases": [ 00:14:07.552 "36601767-4bb2-4ddb-a0fa-de40780f6ce4" 00:14:07.552 ], 00:14:07.552 "product_name": "Malloc disk", 00:14:07.552 "block_size": 512, 00:14:07.552 "num_blocks": 65536, 00:14:07.552 "uuid": "36601767-4bb2-4ddb-a0fa-de40780f6ce4", 00:14:07.552 "assigned_rate_limits": { 00:14:07.552 "rw_ios_per_sec": 0, 00:14:07.552 "rw_mbytes_per_sec": 0, 00:14:07.552 "r_mbytes_per_sec": 0, 00:14:07.552 "w_mbytes_per_sec": 0 00:14:07.552 }, 00:14:07.552 "claimed": true, 00:14:07.552 "claim_type": "exclusive_write", 00:14:07.552 "zoned": false, 00:14:07.552 "supported_io_types": { 00:14:07.552 "read": true, 00:14:07.552 "write": true, 00:14:07.552 "unmap": true, 00:14:07.552 "flush": true, 00:14:07.552 "reset": true, 00:14:07.552 "nvme_admin": false, 00:14:07.552 "nvme_io": false, 00:14:07.552 "nvme_io_md": false, 00:14:07.552 "write_zeroes": true, 00:14:07.552 "zcopy": true, 00:14:07.552 "get_zone_info": false, 00:14:07.552 "zone_management": false, 00:14:07.552 "zone_append": false, 00:14:07.552 "compare": false, 00:14:07.552 "compare_and_write": false, 00:14:07.552 "abort": true, 00:14:07.552 "seek_hole": false, 00:14:07.552 "seek_data": false, 00:14:07.552 "copy": true, 00:14:07.552 "nvme_iov_md": false 00:14:07.552 }, 00:14:07.552 "memory_domains": [ 00:14:07.552 { 00:14:07.552 "dma_device_id": "system", 00:14:07.552 "dma_device_type": 1 00:14:07.552 }, 00:14:07.552 { 00:14:07.552 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:07.552 "dma_device_type": 2 00:14:07.552 } 00:14:07.552 ], 00:14:07.552 "driver_specific": {} 00:14:07.552 } 00:14:07.552 ] 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:07.811 "name": "Existed_Raid", 00:14:07.811 "uuid": "b0394991-535c-4e62-9dd6-43147841e556", 00:14:07.811 "strip_size_kb": 64, 00:14:07.811 "state": "online", 00:14:07.811 "raid_level": "raid5f", 00:14:07.811 "superblock": false, 00:14:07.811 "num_base_bdevs": 4, 00:14:07.811 "num_base_bdevs_discovered": 4, 00:14:07.811 "num_base_bdevs_operational": 4, 00:14:07.811 "base_bdevs_list": [ 00:14:07.811 { 00:14:07.811 "name": "NewBaseBdev", 00:14:07.811 "uuid": "36601767-4bb2-4ddb-a0fa-de40780f6ce4", 00:14:07.811 "is_configured": true, 00:14:07.811 "data_offset": 0, 00:14:07.811 "data_size": 65536 00:14:07.811 }, 00:14:07.811 { 00:14:07.811 "name": "BaseBdev2", 00:14:07.811 "uuid": "ddd7f038-46e7-46ef-b5ec-388222005ef2", 00:14:07.811 "is_configured": true, 00:14:07.811 "data_offset": 0, 00:14:07.811 "data_size": 65536 00:14:07.811 }, 00:14:07.811 { 00:14:07.811 "name": "BaseBdev3", 00:14:07.811 "uuid": "b74ee733-3fdf-447b-a800-76acb45f8c66", 00:14:07.811 "is_configured": true, 00:14:07.811 "data_offset": 0, 00:14:07.811 "data_size": 65536 00:14:07.811 }, 00:14:07.811 { 00:14:07.811 "name": "BaseBdev4", 00:14:07.811 "uuid": "a844086a-c9ce-4f4e-aa85-d68412c7ebd3", 00:14:07.811 "is_configured": true, 00:14:07.811 "data_offset": 0, 00:14:07.811 "data_size": 65536 00:14:07.811 } 00:14:07.811 ] 00:14:07.811 }' 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:07.811 05:41:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.070 [2024-12-07 05:41:41.343871] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:08.070 "name": "Existed_Raid", 00:14:08.070 "aliases": [ 00:14:08.070 "b0394991-535c-4e62-9dd6-43147841e556" 00:14:08.070 ], 00:14:08.070 "product_name": "Raid Volume", 00:14:08.070 "block_size": 512, 00:14:08.070 "num_blocks": 196608, 00:14:08.070 "uuid": "b0394991-535c-4e62-9dd6-43147841e556", 00:14:08.070 "assigned_rate_limits": { 00:14:08.070 "rw_ios_per_sec": 0, 00:14:08.070 "rw_mbytes_per_sec": 0, 00:14:08.070 "r_mbytes_per_sec": 0, 00:14:08.070 "w_mbytes_per_sec": 0 00:14:08.070 }, 00:14:08.070 "claimed": false, 00:14:08.070 "zoned": false, 00:14:08.070 "supported_io_types": { 00:14:08.070 "read": true, 00:14:08.070 "write": true, 00:14:08.070 "unmap": false, 00:14:08.070 "flush": false, 00:14:08.070 "reset": true, 00:14:08.070 "nvme_admin": false, 00:14:08.070 "nvme_io": false, 00:14:08.070 "nvme_io_md": false, 00:14:08.070 "write_zeroes": true, 00:14:08.070 "zcopy": false, 00:14:08.070 "get_zone_info": false, 00:14:08.070 "zone_management": false, 00:14:08.070 "zone_append": false, 00:14:08.070 "compare": false, 00:14:08.070 "compare_and_write": false, 00:14:08.070 "abort": false, 00:14:08.070 "seek_hole": false, 00:14:08.070 "seek_data": false, 00:14:08.070 "copy": false, 00:14:08.070 "nvme_iov_md": false 00:14:08.070 }, 00:14:08.070 "driver_specific": { 00:14:08.070 "raid": { 00:14:08.070 "uuid": "b0394991-535c-4e62-9dd6-43147841e556", 00:14:08.070 "strip_size_kb": 64, 00:14:08.070 "state": "online", 00:14:08.070 "raid_level": "raid5f", 00:14:08.070 "superblock": false, 00:14:08.070 "num_base_bdevs": 4, 00:14:08.070 "num_base_bdevs_discovered": 4, 00:14:08.070 "num_base_bdevs_operational": 4, 00:14:08.070 "base_bdevs_list": [ 00:14:08.070 { 00:14:08.070 "name": "NewBaseBdev", 00:14:08.070 "uuid": "36601767-4bb2-4ddb-a0fa-de40780f6ce4", 00:14:08.070 "is_configured": true, 00:14:08.070 "data_offset": 0, 00:14:08.070 "data_size": 65536 00:14:08.070 }, 00:14:08.070 { 00:14:08.070 "name": "BaseBdev2", 00:14:08.070 "uuid": "ddd7f038-46e7-46ef-b5ec-388222005ef2", 00:14:08.070 "is_configured": true, 00:14:08.070 "data_offset": 0, 00:14:08.070 "data_size": 65536 00:14:08.070 }, 00:14:08.070 { 00:14:08.070 "name": "BaseBdev3", 00:14:08.070 "uuid": "b74ee733-3fdf-447b-a800-76acb45f8c66", 00:14:08.070 "is_configured": true, 00:14:08.070 "data_offset": 0, 00:14:08.070 "data_size": 65536 00:14:08.070 }, 00:14:08.070 { 00:14:08.070 "name": "BaseBdev4", 00:14:08.070 "uuid": "a844086a-c9ce-4f4e-aa85-d68412c7ebd3", 00:14:08.070 "is_configured": true, 00:14:08.070 "data_offset": 0, 00:14:08.070 "data_size": 65536 00:14:08.070 } 00:14:08.070 ] 00:14:08.070 } 00:14:08.070 } 00:14:08.070 }' 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:08.070 BaseBdev2 00:14:08.070 BaseBdev3 00:14:08.070 BaseBdev4' 00:14:08.070 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:08.329 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:08.329 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:08.329 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:08.329 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.330 [2024-12-07 05:41:41.631150] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:08.330 [2024-12-07 05:41:41.631182] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:08.330 [2024-12-07 05:41:41.631257] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:08.330 [2024-12-07 05:41:41.631511] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:08.330 [2024-12-07 05:41:41.631530] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 92846 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 92846 ']' 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 92846 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 92846 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:08.330 killing process with pid 92846 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 92846' 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 92846 00:14:08.330 [2024-12-07 05:41:41.677440] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:08.330 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 92846 00:14:08.589 [2024-12-07 05:41:41.717694] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:08.589 05:41:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:14:08.589 00:14:08.589 real 0m9.111s 00:14:08.589 user 0m15.597s 00:14:08.589 sys 0m1.887s 00:14:08.589 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:08.589 05:41:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.589 ************************************ 00:14:08.589 END TEST raid5f_state_function_test 00:14:08.589 ************************************ 00:14:08.850 05:41:41 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:14:08.850 05:41:41 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:14:08.850 05:41:41 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:08.850 05:41:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:08.850 ************************************ 00:14:08.850 START TEST raid5f_state_function_test_sb 00:14:08.850 ************************************ 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 true 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=93490 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:08.850 Process raid pid: 93490 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 93490' 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 93490 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 93490 ']' 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:08.850 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:08.850 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:08.850 [2024-12-07 05:41:42.100365] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:14:08.850 [2024-12-07 05:41:42.100495] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:09.110 [2024-12-07 05:41:42.235052] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:09.110 [2024-12-07 05:41:42.259977] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:09.110 [2024-12-07 05:41:42.302128] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:09.110 [2024-12-07 05:41:42.302182] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.681 [2024-12-07 05:41:42.940537] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:09.681 [2024-12-07 05:41:42.940711] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:09.681 [2024-12-07 05:41:42.940743] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:09.681 [2024-12-07 05:41:42.940754] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:09.681 [2024-12-07 05:41:42.940760] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:09.681 [2024-12-07 05:41:42.940771] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:09.681 [2024-12-07 05:41:42.940777] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:09.681 [2024-12-07 05:41:42.940786] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:09.681 "name": "Existed_Raid", 00:14:09.681 "uuid": "91267cca-ac3e-4522-bd91-2f34caf3992a", 00:14:09.681 "strip_size_kb": 64, 00:14:09.681 "state": "configuring", 00:14:09.681 "raid_level": "raid5f", 00:14:09.681 "superblock": true, 00:14:09.681 "num_base_bdevs": 4, 00:14:09.681 "num_base_bdevs_discovered": 0, 00:14:09.681 "num_base_bdevs_operational": 4, 00:14:09.681 "base_bdevs_list": [ 00:14:09.681 { 00:14:09.681 "name": "BaseBdev1", 00:14:09.681 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.681 "is_configured": false, 00:14:09.681 "data_offset": 0, 00:14:09.681 "data_size": 0 00:14:09.681 }, 00:14:09.681 { 00:14:09.681 "name": "BaseBdev2", 00:14:09.681 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.681 "is_configured": false, 00:14:09.681 "data_offset": 0, 00:14:09.681 "data_size": 0 00:14:09.681 }, 00:14:09.681 { 00:14:09.681 "name": "BaseBdev3", 00:14:09.681 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.681 "is_configured": false, 00:14:09.681 "data_offset": 0, 00:14:09.681 "data_size": 0 00:14:09.681 }, 00:14:09.681 { 00:14:09.681 "name": "BaseBdev4", 00:14:09.681 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.681 "is_configured": false, 00:14:09.681 "data_offset": 0, 00:14:09.681 "data_size": 0 00:14:09.681 } 00:14:09.681 ] 00:14:09.681 }' 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:09.681 05:41:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.252 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:10.252 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.252 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.252 [2024-12-07 05:41:43.391674] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:10.252 [2024-12-07 05:41:43.391715] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:14:10.252 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.252 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:10.252 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.252 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.252 [2024-12-07 05:41:43.403685] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:10.252 [2024-12-07 05:41:43.403729] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:10.252 [2024-12-07 05:41:43.403738] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:10.252 [2024-12-07 05:41:43.403747] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:10.252 [2024-12-07 05:41:43.403753] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:10.252 [2024-12-07 05:41:43.403762] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:10.252 [2024-12-07 05:41:43.403768] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:10.253 [2024-12-07 05:41:43.403777] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.253 [2024-12-07 05:41:43.424500] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:10.253 BaseBdev1 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.253 [ 00:14:10.253 { 00:14:10.253 "name": "BaseBdev1", 00:14:10.253 "aliases": [ 00:14:10.253 "26640862-c9f4-4d18-836c-09cf0cb5961a" 00:14:10.253 ], 00:14:10.253 "product_name": "Malloc disk", 00:14:10.253 "block_size": 512, 00:14:10.253 "num_blocks": 65536, 00:14:10.253 "uuid": "26640862-c9f4-4d18-836c-09cf0cb5961a", 00:14:10.253 "assigned_rate_limits": { 00:14:10.253 "rw_ios_per_sec": 0, 00:14:10.253 "rw_mbytes_per_sec": 0, 00:14:10.253 "r_mbytes_per_sec": 0, 00:14:10.253 "w_mbytes_per_sec": 0 00:14:10.253 }, 00:14:10.253 "claimed": true, 00:14:10.253 "claim_type": "exclusive_write", 00:14:10.253 "zoned": false, 00:14:10.253 "supported_io_types": { 00:14:10.253 "read": true, 00:14:10.253 "write": true, 00:14:10.253 "unmap": true, 00:14:10.253 "flush": true, 00:14:10.253 "reset": true, 00:14:10.253 "nvme_admin": false, 00:14:10.253 "nvme_io": false, 00:14:10.253 "nvme_io_md": false, 00:14:10.253 "write_zeroes": true, 00:14:10.253 "zcopy": true, 00:14:10.253 "get_zone_info": false, 00:14:10.253 "zone_management": false, 00:14:10.253 "zone_append": false, 00:14:10.253 "compare": false, 00:14:10.253 "compare_and_write": false, 00:14:10.253 "abort": true, 00:14:10.253 "seek_hole": false, 00:14:10.253 "seek_data": false, 00:14:10.253 "copy": true, 00:14:10.253 "nvme_iov_md": false 00:14:10.253 }, 00:14:10.253 "memory_domains": [ 00:14:10.253 { 00:14:10.253 "dma_device_id": "system", 00:14:10.253 "dma_device_type": 1 00:14:10.253 }, 00:14:10.253 { 00:14:10.253 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:10.253 "dma_device_type": 2 00:14:10.253 } 00:14:10.253 ], 00:14:10.253 "driver_specific": {} 00:14:10.253 } 00:14:10.253 ] 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:10.253 "name": "Existed_Raid", 00:14:10.253 "uuid": "96516063-6a25-4a54-8969-e5787f9a486b", 00:14:10.253 "strip_size_kb": 64, 00:14:10.253 "state": "configuring", 00:14:10.253 "raid_level": "raid5f", 00:14:10.253 "superblock": true, 00:14:10.253 "num_base_bdevs": 4, 00:14:10.253 "num_base_bdevs_discovered": 1, 00:14:10.253 "num_base_bdevs_operational": 4, 00:14:10.253 "base_bdevs_list": [ 00:14:10.253 { 00:14:10.253 "name": "BaseBdev1", 00:14:10.253 "uuid": "26640862-c9f4-4d18-836c-09cf0cb5961a", 00:14:10.253 "is_configured": true, 00:14:10.253 "data_offset": 2048, 00:14:10.253 "data_size": 63488 00:14:10.253 }, 00:14:10.253 { 00:14:10.253 "name": "BaseBdev2", 00:14:10.253 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.253 "is_configured": false, 00:14:10.253 "data_offset": 0, 00:14:10.253 "data_size": 0 00:14:10.253 }, 00:14:10.253 { 00:14:10.253 "name": "BaseBdev3", 00:14:10.253 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.253 "is_configured": false, 00:14:10.253 "data_offset": 0, 00:14:10.253 "data_size": 0 00:14:10.253 }, 00:14:10.253 { 00:14:10.253 "name": "BaseBdev4", 00:14:10.253 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.253 "is_configured": false, 00:14:10.253 "data_offset": 0, 00:14:10.253 "data_size": 0 00:14:10.253 } 00:14:10.253 ] 00:14:10.253 }' 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:10.253 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.513 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:10.513 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.513 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.513 [2024-12-07 05:41:43.875745] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:10.513 [2024-12-07 05:41:43.875788] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.773 [2024-12-07 05:41:43.887786] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:10.773 [2024-12-07 05:41:43.889604] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:10.773 [2024-12-07 05:41:43.889653] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:10.773 [2024-12-07 05:41:43.889663] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:10.773 [2024-12-07 05:41:43.889672] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:10.773 [2024-12-07 05:41:43.889678] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:10.773 [2024-12-07 05:41:43.889686] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.773 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:10.773 "name": "Existed_Raid", 00:14:10.773 "uuid": "b7d1ac3b-ec10-4dd8-b44d-4bec0ea9f861", 00:14:10.773 "strip_size_kb": 64, 00:14:10.773 "state": "configuring", 00:14:10.773 "raid_level": "raid5f", 00:14:10.773 "superblock": true, 00:14:10.773 "num_base_bdevs": 4, 00:14:10.773 "num_base_bdevs_discovered": 1, 00:14:10.773 "num_base_bdevs_operational": 4, 00:14:10.773 "base_bdevs_list": [ 00:14:10.773 { 00:14:10.773 "name": "BaseBdev1", 00:14:10.773 "uuid": "26640862-c9f4-4d18-836c-09cf0cb5961a", 00:14:10.773 "is_configured": true, 00:14:10.773 "data_offset": 2048, 00:14:10.773 "data_size": 63488 00:14:10.773 }, 00:14:10.773 { 00:14:10.773 "name": "BaseBdev2", 00:14:10.773 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.773 "is_configured": false, 00:14:10.773 "data_offset": 0, 00:14:10.773 "data_size": 0 00:14:10.773 }, 00:14:10.773 { 00:14:10.773 "name": "BaseBdev3", 00:14:10.773 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.773 "is_configured": false, 00:14:10.773 "data_offset": 0, 00:14:10.773 "data_size": 0 00:14:10.773 }, 00:14:10.773 { 00:14:10.773 "name": "BaseBdev4", 00:14:10.774 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.774 "is_configured": false, 00:14:10.774 "data_offset": 0, 00:14:10.774 "data_size": 0 00:14:10.774 } 00:14:10.774 ] 00:14:10.774 }' 00:14:10.774 05:41:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:10.774 05:41:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.034 [2024-12-07 05:41:44.353823] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:11.034 BaseBdev2 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.034 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.034 [ 00:14:11.034 { 00:14:11.034 "name": "BaseBdev2", 00:14:11.034 "aliases": [ 00:14:11.034 "35ff896b-593f-4de7-ad7d-513776e8d29c" 00:14:11.034 ], 00:14:11.034 "product_name": "Malloc disk", 00:14:11.034 "block_size": 512, 00:14:11.034 "num_blocks": 65536, 00:14:11.035 "uuid": "35ff896b-593f-4de7-ad7d-513776e8d29c", 00:14:11.035 "assigned_rate_limits": { 00:14:11.035 "rw_ios_per_sec": 0, 00:14:11.035 "rw_mbytes_per_sec": 0, 00:14:11.035 "r_mbytes_per_sec": 0, 00:14:11.035 "w_mbytes_per_sec": 0 00:14:11.035 }, 00:14:11.035 "claimed": true, 00:14:11.035 "claim_type": "exclusive_write", 00:14:11.035 "zoned": false, 00:14:11.035 "supported_io_types": { 00:14:11.035 "read": true, 00:14:11.035 "write": true, 00:14:11.035 "unmap": true, 00:14:11.035 "flush": true, 00:14:11.035 "reset": true, 00:14:11.035 "nvme_admin": false, 00:14:11.035 "nvme_io": false, 00:14:11.035 "nvme_io_md": false, 00:14:11.035 "write_zeroes": true, 00:14:11.035 "zcopy": true, 00:14:11.035 "get_zone_info": false, 00:14:11.035 "zone_management": false, 00:14:11.035 "zone_append": false, 00:14:11.035 "compare": false, 00:14:11.035 "compare_and_write": false, 00:14:11.035 "abort": true, 00:14:11.035 "seek_hole": false, 00:14:11.035 "seek_data": false, 00:14:11.035 "copy": true, 00:14:11.035 "nvme_iov_md": false 00:14:11.035 }, 00:14:11.035 "memory_domains": [ 00:14:11.035 { 00:14:11.035 "dma_device_id": "system", 00:14:11.035 "dma_device_type": 1 00:14:11.035 }, 00:14:11.035 { 00:14:11.035 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:11.035 "dma_device_type": 2 00:14:11.035 } 00:14:11.035 ], 00:14:11.035 "driver_specific": {} 00:14:11.035 } 00:14:11.035 ] 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.035 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.295 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.295 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:11.295 "name": "Existed_Raid", 00:14:11.295 "uuid": "b7d1ac3b-ec10-4dd8-b44d-4bec0ea9f861", 00:14:11.295 "strip_size_kb": 64, 00:14:11.295 "state": "configuring", 00:14:11.295 "raid_level": "raid5f", 00:14:11.295 "superblock": true, 00:14:11.295 "num_base_bdevs": 4, 00:14:11.295 "num_base_bdevs_discovered": 2, 00:14:11.295 "num_base_bdevs_operational": 4, 00:14:11.295 "base_bdevs_list": [ 00:14:11.295 { 00:14:11.295 "name": "BaseBdev1", 00:14:11.295 "uuid": "26640862-c9f4-4d18-836c-09cf0cb5961a", 00:14:11.295 "is_configured": true, 00:14:11.295 "data_offset": 2048, 00:14:11.295 "data_size": 63488 00:14:11.295 }, 00:14:11.295 { 00:14:11.295 "name": "BaseBdev2", 00:14:11.295 "uuid": "35ff896b-593f-4de7-ad7d-513776e8d29c", 00:14:11.295 "is_configured": true, 00:14:11.295 "data_offset": 2048, 00:14:11.295 "data_size": 63488 00:14:11.295 }, 00:14:11.295 { 00:14:11.295 "name": "BaseBdev3", 00:14:11.295 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.295 "is_configured": false, 00:14:11.295 "data_offset": 0, 00:14:11.295 "data_size": 0 00:14:11.295 }, 00:14:11.295 { 00:14:11.295 "name": "BaseBdev4", 00:14:11.295 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.295 "is_configured": false, 00:14:11.295 "data_offset": 0, 00:14:11.295 "data_size": 0 00:14:11.295 } 00:14:11.295 ] 00:14:11.295 }' 00:14:11.295 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:11.295 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.555 [2024-12-07 05:41:44.864589] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:11.555 BaseBdev3 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.555 [ 00:14:11.555 { 00:14:11.555 "name": "BaseBdev3", 00:14:11.555 "aliases": [ 00:14:11.555 "3de396b8-1e6b-44f0-95cc-a133e0f99c51" 00:14:11.555 ], 00:14:11.555 "product_name": "Malloc disk", 00:14:11.555 "block_size": 512, 00:14:11.555 "num_blocks": 65536, 00:14:11.555 "uuid": "3de396b8-1e6b-44f0-95cc-a133e0f99c51", 00:14:11.555 "assigned_rate_limits": { 00:14:11.555 "rw_ios_per_sec": 0, 00:14:11.555 "rw_mbytes_per_sec": 0, 00:14:11.555 "r_mbytes_per_sec": 0, 00:14:11.555 "w_mbytes_per_sec": 0 00:14:11.555 }, 00:14:11.555 "claimed": true, 00:14:11.555 "claim_type": "exclusive_write", 00:14:11.555 "zoned": false, 00:14:11.555 "supported_io_types": { 00:14:11.555 "read": true, 00:14:11.555 "write": true, 00:14:11.555 "unmap": true, 00:14:11.555 "flush": true, 00:14:11.555 "reset": true, 00:14:11.555 "nvme_admin": false, 00:14:11.555 "nvme_io": false, 00:14:11.555 "nvme_io_md": false, 00:14:11.555 "write_zeroes": true, 00:14:11.555 "zcopy": true, 00:14:11.555 "get_zone_info": false, 00:14:11.555 "zone_management": false, 00:14:11.555 "zone_append": false, 00:14:11.555 "compare": false, 00:14:11.555 "compare_and_write": false, 00:14:11.555 "abort": true, 00:14:11.555 "seek_hole": false, 00:14:11.555 "seek_data": false, 00:14:11.555 "copy": true, 00:14:11.555 "nvme_iov_md": false 00:14:11.555 }, 00:14:11.555 "memory_domains": [ 00:14:11.555 { 00:14:11.555 "dma_device_id": "system", 00:14:11.555 "dma_device_type": 1 00:14:11.555 }, 00:14:11.555 { 00:14:11.555 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:11.555 "dma_device_type": 2 00:14:11.555 } 00:14:11.555 ], 00:14:11.555 "driver_specific": {} 00:14:11.555 } 00:14:11.555 ] 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.555 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.816 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.816 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:11.816 "name": "Existed_Raid", 00:14:11.816 "uuid": "b7d1ac3b-ec10-4dd8-b44d-4bec0ea9f861", 00:14:11.816 "strip_size_kb": 64, 00:14:11.816 "state": "configuring", 00:14:11.816 "raid_level": "raid5f", 00:14:11.816 "superblock": true, 00:14:11.816 "num_base_bdevs": 4, 00:14:11.816 "num_base_bdevs_discovered": 3, 00:14:11.816 "num_base_bdevs_operational": 4, 00:14:11.816 "base_bdevs_list": [ 00:14:11.816 { 00:14:11.816 "name": "BaseBdev1", 00:14:11.816 "uuid": "26640862-c9f4-4d18-836c-09cf0cb5961a", 00:14:11.816 "is_configured": true, 00:14:11.816 "data_offset": 2048, 00:14:11.816 "data_size": 63488 00:14:11.816 }, 00:14:11.816 { 00:14:11.816 "name": "BaseBdev2", 00:14:11.816 "uuid": "35ff896b-593f-4de7-ad7d-513776e8d29c", 00:14:11.816 "is_configured": true, 00:14:11.816 "data_offset": 2048, 00:14:11.816 "data_size": 63488 00:14:11.816 }, 00:14:11.816 { 00:14:11.816 "name": "BaseBdev3", 00:14:11.816 "uuid": "3de396b8-1e6b-44f0-95cc-a133e0f99c51", 00:14:11.816 "is_configured": true, 00:14:11.816 "data_offset": 2048, 00:14:11.816 "data_size": 63488 00:14:11.816 }, 00:14:11.816 { 00:14:11.816 "name": "BaseBdev4", 00:14:11.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.816 "is_configured": false, 00:14:11.816 "data_offset": 0, 00:14:11.816 "data_size": 0 00:14:11.816 } 00:14:11.816 ] 00:14:11.816 }' 00:14:11.816 05:41:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:11.816 05:41:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.075 [2024-12-07 05:41:45.326730] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:12.075 [2024-12-07 05:41:45.326949] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:12.075 [2024-12-07 05:41:45.326970] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:12.075 [2024-12-07 05:41:45.327267] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:12.075 BaseBdev4 00:14:12.075 [2024-12-07 05:41:45.327790] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:12.075 [2024-12-07 05:41:45.327813] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:14:12.075 [2024-12-07 05:41:45.327951] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.075 [ 00:14:12.075 { 00:14:12.075 "name": "BaseBdev4", 00:14:12.075 "aliases": [ 00:14:12.075 "28ae6e64-5cc3-48e8-b93f-c6c425a5f0a7" 00:14:12.075 ], 00:14:12.075 "product_name": "Malloc disk", 00:14:12.075 "block_size": 512, 00:14:12.075 "num_blocks": 65536, 00:14:12.075 "uuid": "28ae6e64-5cc3-48e8-b93f-c6c425a5f0a7", 00:14:12.075 "assigned_rate_limits": { 00:14:12.075 "rw_ios_per_sec": 0, 00:14:12.075 "rw_mbytes_per_sec": 0, 00:14:12.075 "r_mbytes_per_sec": 0, 00:14:12.075 "w_mbytes_per_sec": 0 00:14:12.075 }, 00:14:12.075 "claimed": true, 00:14:12.075 "claim_type": "exclusive_write", 00:14:12.075 "zoned": false, 00:14:12.075 "supported_io_types": { 00:14:12.075 "read": true, 00:14:12.075 "write": true, 00:14:12.075 "unmap": true, 00:14:12.075 "flush": true, 00:14:12.075 "reset": true, 00:14:12.075 "nvme_admin": false, 00:14:12.075 "nvme_io": false, 00:14:12.075 "nvme_io_md": false, 00:14:12.075 "write_zeroes": true, 00:14:12.075 "zcopy": true, 00:14:12.075 "get_zone_info": false, 00:14:12.075 "zone_management": false, 00:14:12.075 "zone_append": false, 00:14:12.075 "compare": false, 00:14:12.075 "compare_and_write": false, 00:14:12.075 "abort": true, 00:14:12.075 "seek_hole": false, 00:14:12.075 "seek_data": false, 00:14:12.075 "copy": true, 00:14:12.075 "nvme_iov_md": false 00:14:12.075 }, 00:14:12.075 "memory_domains": [ 00:14:12.075 { 00:14:12.075 "dma_device_id": "system", 00:14:12.075 "dma_device_type": 1 00:14:12.075 }, 00:14:12.075 { 00:14:12.075 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:12.075 "dma_device_type": 2 00:14:12.075 } 00:14:12.075 ], 00:14:12.075 "driver_specific": {} 00:14:12.075 } 00:14:12.075 ] 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.075 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.075 "name": "Existed_Raid", 00:14:12.075 "uuid": "b7d1ac3b-ec10-4dd8-b44d-4bec0ea9f861", 00:14:12.075 "strip_size_kb": 64, 00:14:12.075 "state": "online", 00:14:12.075 "raid_level": "raid5f", 00:14:12.075 "superblock": true, 00:14:12.075 "num_base_bdevs": 4, 00:14:12.075 "num_base_bdevs_discovered": 4, 00:14:12.075 "num_base_bdevs_operational": 4, 00:14:12.075 "base_bdevs_list": [ 00:14:12.075 { 00:14:12.075 "name": "BaseBdev1", 00:14:12.075 "uuid": "26640862-c9f4-4d18-836c-09cf0cb5961a", 00:14:12.075 "is_configured": true, 00:14:12.075 "data_offset": 2048, 00:14:12.075 "data_size": 63488 00:14:12.075 }, 00:14:12.075 { 00:14:12.075 "name": "BaseBdev2", 00:14:12.075 "uuid": "35ff896b-593f-4de7-ad7d-513776e8d29c", 00:14:12.075 "is_configured": true, 00:14:12.075 "data_offset": 2048, 00:14:12.075 "data_size": 63488 00:14:12.075 }, 00:14:12.075 { 00:14:12.075 "name": "BaseBdev3", 00:14:12.075 "uuid": "3de396b8-1e6b-44f0-95cc-a133e0f99c51", 00:14:12.075 "is_configured": true, 00:14:12.075 "data_offset": 2048, 00:14:12.075 "data_size": 63488 00:14:12.075 }, 00:14:12.076 { 00:14:12.076 "name": "BaseBdev4", 00:14:12.076 "uuid": "28ae6e64-5cc3-48e8-b93f-c6c425a5f0a7", 00:14:12.076 "is_configured": true, 00:14:12.076 "data_offset": 2048, 00:14:12.076 "data_size": 63488 00:14:12.076 } 00:14:12.076 ] 00:14:12.076 }' 00:14:12.076 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.076 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.644 [2024-12-07 05:41:45.826213] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:12.644 "name": "Existed_Raid", 00:14:12.644 "aliases": [ 00:14:12.644 "b7d1ac3b-ec10-4dd8-b44d-4bec0ea9f861" 00:14:12.644 ], 00:14:12.644 "product_name": "Raid Volume", 00:14:12.644 "block_size": 512, 00:14:12.644 "num_blocks": 190464, 00:14:12.644 "uuid": "b7d1ac3b-ec10-4dd8-b44d-4bec0ea9f861", 00:14:12.644 "assigned_rate_limits": { 00:14:12.644 "rw_ios_per_sec": 0, 00:14:12.644 "rw_mbytes_per_sec": 0, 00:14:12.644 "r_mbytes_per_sec": 0, 00:14:12.644 "w_mbytes_per_sec": 0 00:14:12.644 }, 00:14:12.644 "claimed": false, 00:14:12.644 "zoned": false, 00:14:12.644 "supported_io_types": { 00:14:12.644 "read": true, 00:14:12.644 "write": true, 00:14:12.644 "unmap": false, 00:14:12.644 "flush": false, 00:14:12.644 "reset": true, 00:14:12.644 "nvme_admin": false, 00:14:12.644 "nvme_io": false, 00:14:12.644 "nvme_io_md": false, 00:14:12.644 "write_zeroes": true, 00:14:12.644 "zcopy": false, 00:14:12.644 "get_zone_info": false, 00:14:12.644 "zone_management": false, 00:14:12.644 "zone_append": false, 00:14:12.644 "compare": false, 00:14:12.644 "compare_and_write": false, 00:14:12.644 "abort": false, 00:14:12.644 "seek_hole": false, 00:14:12.644 "seek_data": false, 00:14:12.644 "copy": false, 00:14:12.644 "nvme_iov_md": false 00:14:12.644 }, 00:14:12.644 "driver_specific": { 00:14:12.644 "raid": { 00:14:12.644 "uuid": "b7d1ac3b-ec10-4dd8-b44d-4bec0ea9f861", 00:14:12.644 "strip_size_kb": 64, 00:14:12.644 "state": "online", 00:14:12.644 "raid_level": "raid5f", 00:14:12.644 "superblock": true, 00:14:12.644 "num_base_bdevs": 4, 00:14:12.644 "num_base_bdevs_discovered": 4, 00:14:12.644 "num_base_bdevs_operational": 4, 00:14:12.644 "base_bdevs_list": [ 00:14:12.644 { 00:14:12.644 "name": "BaseBdev1", 00:14:12.644 "uuid": "26640862-c9f4-4d18-836c-09cf0cb5961a", 00:14:12.644 "is_configured": true, 00:14:12.644 "data_offset": 2048, 00:14:12.644 "data_size": 63488 00:14:12.644 }, 00:14:12.644 { 00:14:12.644 "name": "BaseBdev2", 00:14:12.644 "uuid": "35ff896b-593f-4de7-ad7d-513776e8d29c", 00:14:12.644 "is_configured": true, 00:14:12.644 "data_offset": 2048, 00:14:12.644 "data_size": 63488 00:14:12.644 }, 00:14:12.644 { 00:14:12.644 "name": "BaseBdev3", 00:14:12.644 "uuid": "3de396b8-1e6b-44f0-95cc-a133e0f99c51", 00:14:12.644 "is_configured": true, 00:14:12.644 "data_offset": 2048, 00:14:12.644 "data_size": 63488 00:14:12.644 }, 00:14:12.644 { 00:14:12.644 "name": "BaseBdev4", 00:14:12.644 "uuid": "28ae6e64-5cc3-48e8-b93f-c6c425a5f0a7", 00:14:12.644 "is_configured": true, 00:14:12.644 "data_offset": 2048, 00:14:12.644 "data_size": 63488 00:14:12.644 } 00:14:12.644 ] 00:14:12.644 } 00:14:12.644 } 00:14:12.644 }' 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:12.644 BaseBdev2 00:14:12.644 BaseBdev3 00:14:12.644 BaseBdev4' 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:12.644 05:41:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:12.644 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.644 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.904 [2024-12-07 05:41:46.097568] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.904 "name": "Existed_Raid", 00:14:12.904 "uuid": "b7d1ac3b-ec10-4dd8-b44d-4bec0ea9f861", 00:14:12.904 "strip_size_kb": 64, 00:14:12.904 "state": "online", 00:14:12.904 "raid_level": "raid5f", 00:14:12.904 "superblock": true, 00:14:12.904 "num_base_bdevs": 4, 00:14:12.904 "num_base_bdevs_discovered": 3, 00:14:12.904 "num_base_bdevs_operational": 3, 00:14:12.904 "base_bdevs_list": [ 00:14:12.904 { 00:14:12.904 "name": null, 00:14:12.904 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.904 "is_configured": false, 00:14:12.904 "data_offset": 0, 00:14:12.904 "data_size": 63488 00:14:12.904 }, 00:14:12.904 { 00:14:12.904 "name": "BaseBdev2", 00:14:12.904 "uuid": "35ff896b-593f-4de7-ad7d-513776e8d29c", 00:14:12.904 "is_configured": true, 00:14:12.904 "data_offset": 2048, 00:14:12.904 "data_size": 63488 00:14:12.904 }, 00:14:12.904 { 00:14:12.904 "name": "BaseBdev3", 00:14:12.904 "uuid": "3de396b8-1e6b-44f0-95cc-a133e0f99c51", 00:14:12.904 "is_configured": true, 00:14:12.904 "data_offset": 2048, 00:14:12.904 "data_size": 63488 00:14:12.904 }, 00:14:12.904 { 00:14:12.904 "name": "BaseBdev4", 00:14:12.904 "uuid": "28ae6e64-5cc3-48e8-b93f-c6c425a5f0a7", 00:14:12.904 "is_configured": true, 00:14:12.904 "data_offset": 2048, 00:14:12.904 "data_size": 63488 00:14:12.904 } 00:14:12.904 ] 00:14:12.904 }' 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.904 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.475 [2024-12-07 05:41:46.599870] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:13.475 [2024-12-07 05:41:46.600014] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:13.475 [2024-12-07 05:41:46.611148] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.475 [2024-12-07 05:41:46.651090] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:13.475 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.476 [2024-12-07 05:41:46.722080] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:13.476 [2024-12-07 05:41:46.722126] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.476 BaseBdev2 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.476 [ 00:14:13.476 { 00:14:13.476 "name": "BaseBdev2", 00:14:13.476 "aliases": [ 00:14:13.476 "afc5c893-7a8a-48e2-b1b9-af155532c0eb" 00:14:13.476 ], 00:14:13.476 "product_name": "Malloc disk", 00:14:13.476 "block_size": 512, 00:14:13.476 "num_blocks": 65536, 00:14:13.476 "uuid": "afc5c893-7a8a-48e2-b1b9-af155532c0eb", 00:14:13.476 "assigned_rate_limits": { 00:14:13.476 "rw_ios_per_sec": 0, 00:14:13.476 "rw_mbytes_per_sec": 0, 00:14:13.476 "r_mbytes_per_sec": 0, 00:14:13.476 "w_mbytes_per_sec": 0 00:14:13.476 }, 00:14:13.476 "claimed": false, 00:14:13.476 "zoned": false, 00:14:13.476 "supported_io_types": { 00:14:13.476 "read": true, 00:14:13.476 "write": true, 00:14:13.476 "unmap": true, 00:14:13.476 "flush": true, 00:14:13.476 "reset": true, 00:14:13.476 "nvme_admin": false, 00:14:13.476 "nvme_io": false, 00:14:13.476 "nvme_io_md": false, 00:14:13.476 "write_zeroes": true, 00:14:13.476 "zcopy": true, 00:14:13.476 "get_zone_info": false, 00:14:13.476 "zone_management": false, 00:14:13.476 "zone_append": false, 00:14:13.476 "compare": false, 00:14:13.476 "compare_and_write": false, 00:14:13.476 "abort": true, 00:14:13.476 "seek_hole": false, 00:14:13.476 "seek_data": false, 00:14:13.476 "copy": true, 00:14:13.476 "nvme_iov_md": false 00:14:13.476 }, 00:14:13.476 "memory_domains": [ 00:14:13.476 { 00:14:13.476 "dma_device_id": "system", 00:14:13.476 "dma_device_type": 1 00:14:13.476 }, 00:14:13.476 { 00:14:13.476 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:13.476 "dma_device_type": 2 00:14:13.476 } 00:14:13.476 ], 00:14:13.476 "driver_specific": {} 00:14:13.476 } 00:14:13.476 ] 00:14:13.476 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.737 BaseBdev3 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.737 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.737 [ 00:14:13.737 { 00:14:13.737 "name": "BaseBdev3", 00:14:13.737 "aliases": [ 00:14:13.737 "154206a3-5647-4582-b9e9-980956cd820e" 00:14:13.737 ], 00:14:13.737 "product_name": "Malloc disk", 00:14:13.737 "block_size": 512, 00:14:13.737 "num_blocks": 65536, 00:14:13.737 "uuid": "154206a3-5647-4582-b9e9-980956cd820e", 00:14:13.737 "assigned_rate_limits": { 00:14:13.737 "rw_ios_per_sec": 0, 00:14:13.737 "rw_mbytes_per_sec": 0, 00:14:13.737 "r_mbytes_per_sec": 0, 00:14:13.737 "w_mbytes_per_sec": 0 00:14:13.737 }, 00:14:13.737 "claimed": false, 00:14:13.738 "zoned": false, 00:14:13.738 "supported_io_types": { 00:14:13.738 "read": true, 00:14:13.738 "write": true, 00:14:13.738 "unmap": true, 00:14:13.738 "flush": true, 00:14:13.738 "reset": true, 00:14:13.738 "nvme_admin": false, 00:14:13.738 "nvme_io": false, 00:14:13.738 "nvme_io_md": false, 00:14:13.738 "write_zeroes": true, 00:14:13.738 "zcopy": true, 00:14:13.738 "get_zone_info": false, 00:14:13.738 "zone_management": false, 00:14:13.738 "zone_append": false, 00:14:13.738 "compare": false, 00:14:13.738 "compare_and_write": false, 00:14:13.738 "abort": true, 00:14:13.738 "seek_hole": false, 00:14:13.738 "seek_data": false, 00:14:13.738 "copy": true, 00:14:13.738 "nvme_iov_md": false 00:14:13.738 }, 00:14:13.738 "memory_domains": [ 00:14:13.738 { 00:14:13.738 "dma_device_id": "system", 00:14:13.738 "dma_device_type": 1 00:14:13.738 }, 00:14:13.738 { 00:14:13.738 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:13.738 "dma_device_type": 2 00:14:13.738 } 00:14:13.738 ], 00:14:13.738 "driver_specific": {} 00:14:13.738 } 00:14:13.738 ] 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.738 BaseBdev4 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.738 [ 00:14:13.738 { 00:14:13.738 "name": "BaseBdev4", 00:14:13.738 "aliases": [ 00:14:13.738 "9dd49a41-461f-4807-847e-65f3edfe320d" 00:14:13.738 ], 00:14:13.738 "product_name": "Malloc disk", 00:14:13.738 "block_size": 512, 00:14:13.738 "num_blocks": 65536, 00:14:13.738 "uuid": "9dd49a41-461f-4807-847e-65f3edfe320d", 00:14:13.738 "assigned_rate_limits": { 00:14:13.738 "rw_ios_per_sec": 0, 00:14:13.738 "rw_mbytes_per_sec": 0, 00:14:13.738 "r_mbytes_per_sec": 0, 00:14:13.738 "w_mbytes_per_sec": 0 00:14:13.738 }, 00:14:13.738 "claimed": false, 00:14:13.738 "zoned": false, 00:14:13.738 "supported_io_types": { 00:14:13.738 "read": true, 00:14:13.738 "write": true, 00:14:13.738 "unmap": true, 00:14:13.738 "flush": true, 00:14:13.738 "reset": true, 00:14:13.738 "nvme_admin": false, 00:14:13.738 "nvme_io": false, 00:14:13.738 "nvme_io_md": false, 00:14:13.738 "write_zeroes": true, 00:14:13.738 "zcopy": true, 00:14:13.738 "get_zone_info": false, 00:14:13.738 "zone_management": false, 00:14:13.738 "zone_append": false, 00:14:13.738 "compare": false, 00:14:13.738 "compare_and_write": false, 00:14:13.738 "abort": true, 00:14:13.738 "seek_hole": false, 00:14:13.738 "seek_data": false, 00:14:13.738 "copy": true, 00:14:13.738 "nvme_iov_md": false 00:14:13.738 }, 00:14:13.738 "memory_domains": [ 00:14:13.738 { 00:14:13.738 "dma_device_id": "system", 00:14:13.738 "dma_device_type": 1 00:14:13.738 }, 00:14:13.738 { 00:14:13.738 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:13.738 "dma_device_type": 2 00:14:13.738 } 00:14:13.738 ], 00:14:13.738 "driver_specific": {} 00:14:13.738 } 00:14:13.738 ] 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.738 [2024-12-07 05:41:46.957056] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:13.738 [2024-12-07 05:41:46.957145] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:13.738 [2024-12-07 05:41:46.957192] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:13.738 [2024-12-07 05:41:46.959032] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:13.738 [2024-12-07 05:41:46.959128] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.738 05:41:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.738 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:13.738 "name": "Existed_Raid", 00:14:13.738 "uuid": "7855c5ce-df38-4467-95df-146d3f9a71ce", 00:14:13.738 "strip_size_kb": 64, 00:14:13.738 "state": "configuring", 00:14:13.738 "raid_level": "raid5f", 00:14:13.738 "superblock": true, 00:14:13.738 "num_base_bdevs": 4, 00:14:13.738 "num_base_bdevs_discovered": 3, 00:14:13.738 "num_base_bdevs_operational": 4, 00:14:13.738 "base_bdevs_list": [ 00:14:13.738 { 00:14:13.738 "name": "BaseBdev1", 00:14:13.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:13.738 "is_configured": false, 00:14:13.738 "data_offset": 0, 00:14:13.738 "data_size": 0 00:14:13.738 }, 00:14:13.738 { 00:14:13.738 "name": "BaseBdev2", 00:14:13.738 "uuid": "afc5c893-7a8a-48e2-b1b9-af155532c0eb", 00:14:13.738 "is_configured": true, 00:14:13.738 "data_offset": 2048, 00:14:13.738 "data_size": 63488 00:14:13.738 }, 00:14:13.738 { 00:14:13.738 "name": "BaseBdev3", 00:14:13.738 "uuid": "154206a3-5647-4582-b9e9-980956cd820e", 00:14:13.738 "is_configured": true, 00:14:13.738 "data_offset": 2048, 00:14:13.738 "data_size": 63488 00:14:13.738 }, 00:14:13.738 { 00:14:13.738 "name": "BaseBdev4", 00:14:13.738 "uuid": "9dd49a41-461f-4807-847e-65f3edfe320d", 00:14:13.738 "is_configured": true, 00:14:13.738 "data_offset": 2048, 00:14:13.738 "data_size": 63488 00:14:13.738 } 00:14:13.738 ] 00:14:13.738 }' 00:14:13.738 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:13.738 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.307 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:14.307 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.307 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.307 [2024-12-07 05:41:47.416218] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:14.307 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.307 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:14.307 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:14.308 "name": "Existed_Raid", 00:14:14.308 "uuid": "7855c5ce-df38-4467-95df-146d3f9a71ce", 00:14:14.308 "strip_size_kb": 64, 00:14:14.308 "state": "configuring", 00:14:14.308 "raid_level": "raid5f", 00:14:14.308 "superblock": true, 00:14:14.308 "num_base_bdevs": 4, 00:14:14.308 "num_base_bdevs_discovered": 2, 00:14:14.308 "num_base_bdevs_operational": 4, 00:14:14.308 "base_bdevs_list": [ 00:14:14.308 { 00:14:14.308 "name": "BaseBdev1", 00:14:14.308 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:14.308 "is_configured": false, 00:14:14.308 "data_offset": 0, 00:14:14.308 "data_size": 0 00:14:14.308 }, 00:14:14.308 { 00:14:14.308 "name": null, 00:14:14.308 "uuid": "afc5c893-7a8a-48e2-b1b9-af155532c0eb", 00:14:14.308 "is_configured": false, 00:14:14.308 "data_offset": 0, 00:14:14.308 "data_size": 63488 00:14:14.308 }, 00:14:14.308 { 00:14:14.308 "name": "BaseBdev3", 00:14:14.308 "uuid": "154206a3-5647-4582-b9e9-980956cd820e", 00:14:14.308 "is_configured": true, 00:14:14.308 "data_offset": 2048, 00:14:14.308 "data_size": 63488 00:14:14.308 }, 00:14:14.308 { 00:14:14.308 "name": "BaseBdev4", 00:14:14.308 "uuid": "9dd49a41-461f-4807-847e-65f3edfe320d", 00:14:14.308 "is_configured": true, 00:14:14.308 "data_offset": 2048, 00:14:14.308 "data_size": 63488 00:14:14.308 } 00:14:14.308 ] 00:14:14.308 }' 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:14.308 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.568 [2024-12-07 05:41:47.870302] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:14.568 BaseBdev1 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.568 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.568 [ 00:14:14.568 { 00:14:14.568 "name": "BaseBdev1", 00:14:14.568 "aliases": [ 00:14:14.568 "a855bcc3-e4e7-4e31-8516-df34dcae7610" 00:14:14.568 ], 00:14:14.568 "product_name": "Malloc disk", 00:14:14.568 "block_size": 512, 00:14:14.568 "num_blocks": 65536, 00:14:14.568 "uuid": "a855bcc3-e4e7-4e31-8516-df34dcae7610", 00:14:14.568 "assigned_rate_limits": { 00:14:14.568 "rw_ios_per_sec": 0, 00:14:14.568 "rw_mbytes_per_sec": 0, 00:14:14.568 "r_mbytes_per_sec": 0, 00:14:14.568 "w_mbytes_per_sec": 0 00:14:14.568 }, 00:14:14.568 "claimed": true, 00:14:14.568 "claim_type": "exclusive_write", 00:14:14.568 "zoned": false, 00:14:14.568 "supported_io_types": { 00:14:14.568 "read": true, 00:14:14.568 "write": true, 00:14:14.568 "unmap": true, 00:14:14.568 "flush": true, 00:14:14.568 "reset": true, 00:14:14.568 "nvme_admin": false, 00:14:14.568 "nvme_io": false, 00:14:14.568 "nvme_io_md": false, 00:14:14.568 "write_zeroes": true, 00:14:14.568 "zcopy": true, 00:14:14.569 "get_zone_info": false, 00:14:14.569 "zone_management": false, 00:14:14.569 "zone_append": false, 00:14:14.569 "compare": false, 00:14:14.569 "compare_and_write": false, 00:14:14.569 "abort": true, 00:14:14.569 "seek_hole": false, 00:14:14.569 "seek_data": false, 00:14:14.569 "copy": true, 00:14:14.569 "nvme_iov_md": false 00:14:14.569 }, 00:14:14.569 "memory_domains": [ 00:14:14.569 { 00:14:14.569 "dma_device_id": "system", 00:14:14.569 "dma_device_type": 1 00:14:14.569 }, 00:14:14.569 { 00:14:14.569 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:14.569 "dma_device_type": 2 00:14:14.569 } 00:14:14.569 ], 00:14:14.569 "driver_specific": {} 00:14:14.569 } 00:14:14.569 ] 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.569 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.828 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.828 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:14.828 "name": "Existed_Raid", 00:14:14.828 "uuid": "7855c5ce-df38-4467-95df-146d3f9a71ce", 00:14:14.828 "strip_size_kb": 64, 00:14:14.828 "state": "configuring", 00:14:14.828 "raid_level": "raid5f", 00:14:14.828 "superblock": true, 00:14:14.828 "num_base_bdevs": 4, 00:14:14.829 "num_base_bdevs_discovered": 3, 00:14:14.829 "num_base_bdevs_operational": 4, 00:14:14.829 "base_bdevs_list": [ 00:14:14.829 { 00:14:14.829 "name": "BaseBdev1", 00:14:14.829 "uuid": "a855bcc3-e4e7-4e31-8516-df34dcae7610", 00:14:14.829 "is_configured": true, 00:14:14.829 "data_offset": 2048, 00:14:14.829 "data_size": 63488 00:14:14.829 }, 00:14:14.829 { 00:14:14.829 "name": null, 00:14:14.829 "uuid": "afc5c893-7a8a-48e2-b1b9-af155532c0eb", 00:14:14.829 "is_configured": false, 00:14:14.829 "data_offset": 0, 00:14:14.829 "data_size": 63488 00:14:14.829 }, 00:14:14.829 { 00:14:14.829 "name": "BaseBdev3", 00:14:14.829 "uuid": "154206a3-5647-4582-b9e9-980956cd820e", 00:14:14.829 "is_configured": true, 00:14:14.829 "data_offset": 2048, 00:14:14.829 "data_size": 63488 00:14:14.829 }, 00:14:14.829 { 00:14:14.829 "name": "BaseBdev4", 00:14:14.829 "uuid": "9dd49a41-461f-4807-847e-65f3edfe320d", 00:14:14.829 "is_configured": true, 00:14:14.829 "data_offset": 2048, 00:14:14.829 "data_size": 63488 00:14:14.829 } 00:14:14.829 ] 00:14:14.829 }' 00:14:14.829 05:41:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:14.829 05:41:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.089 [2024-12-07 05:41:48.377468] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:15.089 "name": "Existed_Raid", 00:14:15.089 "uuid": "7855c5ce-df38-4467-95df-146d3f9a71ce", 00:14:15.089 "strip_size_kb": 64, 00:14:15.089 "state": "configuring", 00:14:15.089 "raid_level": "raid5f", 00:14:15.089 "superblock": true, 00:14:15.089 "num_base_bdevs": 4, 00:14:15.089 "num_base_bdevs_discovered": 2, 00:14:15.089 "num_base_bdevs_operational": 4, 00:14:15.089 "base_bdevs_list": [ 00:14:15.089 { 00:14:15.089 "name": "BaseBdev1", 00:14:15.089 "uuid": "a855bcc3-e4e7-4e31-8516-df34dcae7610", 00:14:15.089 "is_configured": true, 00:14:15.089 "data_offset": 2048, 00:14:15.089 "data_size": 63488 00:14:15.089 }, 00:14:15.089 { 00:14:15.089 "name": null, 00:14:15.089 "uuid": "afc5c893-7a8a-48e2-b1b9-af155532c0eb", 00:14:15.089 "is_configured": false, 00:14:15.089 "data_offset": 0, 00:14:15.089 "data_size": 63488 00:14:15.089 }, 00:14:15.089 { 00:14:15.089 "name": null, 00:14:15.089 "uuid": "154206a3-5647-4582-b9e9-980956cd820e", 00:14:15.089 "is_configured": false, 00:14:15.089 "data_offset": 0, 00:14:15.089 "data_size": 63488 00:14:15.089 }, 00:14:15.089 { 00:14:15.089 "name": "BaseBdev4", 00:14:15.089 "uuid": "9dd49a41-461f-4807-847e-65f3edfe320d", 00:14:15.089 "is_configured": true, 00:14:15.089 "data_offset": 2048, 00:14:15.089 "data_size": 63488 00:14:15.089 } 00:14:15.089 ] 00:14:15.089 }' 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:15.089 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.659 [2024-12-07 05:41:48.884607] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:15.659 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.660 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.660 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.660 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.660 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:15.660 "name": "Existed_Raid", 00:14:15.660 "uuid": "7855c5ce-df38-4467-95df-146d3f9a71ce", 00:14:15.660 "strip_size_kb": 64, 00:14:15.660 "state": "configuring", 00:14:15.660 "raid_level": "raid5f", 00:14:15.660 "superblock": true, 00:14:15.660 "num_base_bdevs": 4, 00:14:15.660 "num_base_bdevs_discovered": 3, 00:14:15.660 "num_base_bdevs_operational": 4, 00:14:15.660 "base_bdevs_list": [ 00:14:15.660 { 00:14:15.660 "name": "BaseBdev1", 00:14:15.660 "uuid": "a855bcc3-e4e7-4e31-8516-df34dcae7610", 00:14:15.660 "is_configured": true, 00:14:15.660 "data_offset": 2048, 00:14:15.660 "data_size": 63488 00:14:15.660 }, 00:14:15.660 { 00:14:15.660 "name": null, 00:14:15.660 "uuid": "afc5c893-7a8a-48e2-b1b9-af155532c0eb", 00:14:15.660 "is_configured": false, 00:14:15.660 "data_offset": 0, 00:14:15.660 "data_size": 63488 00:14:15.660 }, 00:14:15.660 { 00:14:15.660 "name": "BaseBdev3", 00:14:15.660 "uuid": "154206a3-5647-4582-b9e9-980956cd820e", 00:14:15.660 "is_configured": true, 00:14:15.660 "data_offset": 2048, 00:14:15.660 "data_size": 63488 00:14:15.660 }, 00:14:15.660 { 00:14:15.660 "name": "BaseBdev4", 00:14:15.660 "uuid": "9dd49a41-461f-4807-847e-65f3edfe320d", 00:14:15.660 "is_configured": true, 00:14:15.660 "data_offset": 2048, 00:14:15.660 "data_size": 63488 00:14:15.660 } 00:14:15.660 ] 00:14:15.660 }' 00:14:15.660 05:41:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:15.660 05:41:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.919 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.919 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.919 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.919 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:15.919 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.179 [2024-12-07 05:41:49.319900] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:16.179 "name": "Existed_Raid", 00:14:16.179 "uuid": "7855c5ce-df38-4467-95df-146d3f9a71ce", 00:14:16.179 "strip_size_kb": 64, 00:14:16.179 "state": "configuring", 00:14:16.179 "raid_level": "raid5f", 00:14:16.179 "superblock": true, 00:14:16.179 "num_base_bdevs": 4, 00:14:16.179 "num_base_bdevs_discovered": 2, 00:14:16.179 "num_base_bdevs_operational": 4, 00:14:16.179 "base_bdevs_list": [ 00:14:16.179 { 00:14:16.179 "name": null, 00:14:16.179 "uuid": "a855bcc3-e4e7-4e31-8516-df34dcae7610", 00:14:16.179 "is_configured": false, 00:14:16.179 "data_offset": 0, 00:14:16.179 "data_size": 63488 00:14:16.179 }, 00:14:16.179 { 00:14:16.179 "name": null, 00:14:16.179 "uuid": "afc5c893-7a8a-48e2-b1b9-af155532c0eb", 00:14:16.179 "is_configured": false, 00:14:16.179 "data_offset": 0, 00:14:16.179 "data_size": 63488 00:14:16.179 }, 00:14:16.179 { 00:14:16.179 "name": "BaseBdev3", 00:14:16.179 "uuid": "154206a3-5647-4582-b9e9-980956cd820e", 00:14:16.179 "is_configured": true, 00:14:16.179 "data_offset": 2048, 00:14:16.179 "data_size": 63488 00:14:16.179 }, 00:14:16.179 { 00:14:16.179 "name": "BaseBdev4", 00:14:16.179 "uuid": "9dd49a41-461f-4807-847e-65f3edfe320d", 00:14:16.179 "is_configured": true, 00:14:16.179 "data_offset": 2048, 00:14:16.179 "data_size": 63488 00:14:16.179 } 00:14:16.179 ] 00:14:16.179 }' 00:14:16.179 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:16.180 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.452 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:16.452 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.453 [2024-12-07 05:41:49.793467] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.453 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:16.741 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:16.741 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:16.741 "name": "Existed_Raid", 00:14:16.741 "uuid": "7855c5ce-df38-4467-95df-146d3f9a71ce", 00:14:16.741 "strip_size_kb": 64, 00:14:16.741 "state": "configuring", 00:14:16.741 "raid_level": "raid5f", 00:14:16.741 "superblock": true, 00:14:16.741 "num_base_bdevs": 4, 00:14:16.741 "num_base_bdevs_discovered": 3, 00:14:16.741 "num_base_bdevs_operational": 4, 00:14:16.741 "base_bdevs_list": [ 00:14:16.741 { 00:14:16.741 "name": null, 00:14:16.741 "uuid": "a855bcc3-e4e7-4e31-8516-df34dcae7610", 00:14:16.741 "is_configured": false, 00:14:16.741 "data_offset": 0, 00:14:16.741 "data_size": 63488 00:14:16.741 }, 00:14:16.741 { 00:14:16.741 "name": "BaseBdev2", 00:14:16.741 "uuid": "afc5c893-7a8a-48e2-b1b9-af155532c0eb", 00:14:16.741 "is_configured": true, 00:14:16.741 "data_offset": 2048, 00:14:16.741 "data_size": 63488 00:14:16.741 }, 00:14:16.741 { 00:14:16.741 "name": "BaseBdev3", 00:14:16.741 "uuid": "154206a3-5647-4582-b9e9-980956cd820e", 00:14:16.741 "is_configured": true, 00:14:16.741 "data_offset": 2048, 00:14:16.741 "data_size": 63488 00:14:16.741 }, 00:14:16.741 { 00:14:16.741 "name": "BaseBdev4", 00:14:16.741 "uuid": "9dd49a41-461f-4807-847e-65f3edfe320d", 00:14:16.741 "is_configured": true, 00:14:16.741 "data_offset": 2048, 00:14:16.741 "data_size": 63488 00:14:16.741 } 00:14:16.741 ] 00:14:16.741 }' 00:14:16.741 05:41:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:16.742 05:41:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u a855bcc3-e4e7-4e31-8516-df34dcae7610 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.027 [2024-12-07 05:41:50.287460] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:17.027 [2024-12-07 05:41:50.287733] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:17.027 [2024-12-07 05:41:50.287751] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:17.027 [2024-12-07 05:41:50.288005] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:14:17.027 NewBaseBdev 00:14:17.027 [2024-12-07 05:41:50.288467] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:17.027 [2024-12-07 05:41:50.288488] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:14:17.027 [2024-12-07 05:41:50.288586] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.027 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.027 [ 00:14:17.027 { 00:14:17.027 "name": "NewBaseBdev", 00:14:17.027 "aliases": [ 00:14:17.027 "a855bcc3-e4e7-4e31-8516-df34dcae7610" 00:14:17.027 ], 00:14:17.027 "product_name": "Malloc disk", 00:14:17.027 "block_size": 512, 00:14:17.027 "num_blocks": 65536, 00:14:17.027 "uuid": "a855bcc3-e4e7-4e31-8516-df34dcae7610", 00:14:17.027 "assigned_rate_limits": { 00:14:17.027 "rw_ios_per_sec": 0, 00:14:17.027 "rw_mbytes_per_sec": 0, 00:14:17.027 "r_mbytes_per_sec": 0, 00:14:17.027 "w_mbytes_per_sec": 0 00:14:17.027 }, 00:14:17.027 "claimed": true, 00:14:17.027 "claim_type": "exclusive_write", 00:14:17.027 "zoned": false, 00:14:17.027 "supported_io_types": { 00:14:17.027 "read": true, 00:14:17.027 "write": true, 00:14:17.027 "unmap": true, 00:14:17.027 "flush": true, 00:14:17.027 "reset": true, 00:14:17.027 "nvme_admin": false, 00:14:17.027 "nvme_io": false, 00:14:17.027 "nvme_io_md": false, 00:14:17.027 "write_zeroes": true, 00:14:17.027 "zcopy": true, 00:14:17.027 "get_zone_info": false, 00:14:17.027 "zone_management": false, 00:14:17.027 "zone_append": false, 00:14:17.027 "compare": false, 00:14:17.027 "compare_and_write": false, 00:14:17.028 "abort": true, 00:14:17.028 "seek_hole": false, 00:14:17.028 "seek_data": false, 00:14:17.028 "copy": true, 00:14:17.028 "nvme_iov_md": false 00:14:17.028 }, 00:14:17.028 "memory_domains": [ 00:14:17.028 { 00:14:17.028 "dma_device_id": "system", 00:14:17.028 "dma_device_type": 1 00:14:17.028 }, 00:14:17.028 { 00:14:17.028 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:17.028 "dma_device_type": 2 00:14:17.028 } 00:14:17.028 ], 00:14:17.028 "driver_specific": {} 00:14:17.028 } 00:14:17.028 ] 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:17.028 "name": "Existed_Raid", 00:14:17.028 "uuid": "7855c5ce-df38-4467-95df-146d3f9a71ce", 00:14:17.028 "strip_size_kb": 64, 00:14:17.028 "state": "online", 00:14:17.028 "raid_level": "raid5f", 00:14:17.028 "superblock": true, 00:14:17.028 "num_base_bdevs": 4, 00:14:17.028 "num_base_bdevs_discovered": 4, 00:14:17.028 "num_base_bdevs_operational": 4, 00:14:17.028 "base_bdevs_list": [ 00:14:17.028 { 00:14:17.028 "name": "NewBaseBdev", 00:14:17.028 "uuid": "a855bcc3-e4e7-4e31-8516-df34dcae7610", 00:14:17.028 "is_configured": true, 00:14:17.028 "data_offset": 2048, 00:14:17.028 "data_size": 63488 00:14:17.028 }, 00:14:17.028 { 00:14:17.028 "name": "BaseBdev2", 00:14:17.028 "uuid": "afc5c893-7a8a-48e2-b1b9-af155532c0eb", 00:14:17.028 "is_configured": true, 00:14:17.028 "data_offset": 2048, 00:14:17.028 "data_size": 63488 00:14:17.028 }, 00:14:17.028 { 00:14:17.028 "name": "BaseBdev3", 00:14:17.028 "uuid": "154206a3-5647-4582-b9e9-980956cd820e", 00:14:17.028 "is_configured": true, 00:14:17.028 "data_offset": 2048, 00:14:17.028 "data_size": 63488 00:14:17.028 }, 00:14:17.028 { 00:14:17.028 "name": "BaseBdev4", 00:14:17.028 "uuid": "9dd49a41-461f-4807-847e-65f3edfe320d", 00:14:17.028 "is_configured": true, 00:14:17.028 "data_offset": 2048, 00:14:17.028 "data_size": 63488 00:14:17.028 } 00:14:17.028 ] 00:14:17.028 }' 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:17.028 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.609 [2024-12-07 05:41:50.766865] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:17.609 "name": "Existed_Raid", 00:14:17.609 "aliases": [ 00:14:17.609 "7855c5ce-df38-4467-95df-146d3f9a71ce" 00:14:17.609 ], 00:14:17.609 "product_name": "Raid Volume", 00:14:17.609 "block_size": 512, 00:14:17.609 "num_blocks": 190464, 00:14:17.609 "uuid": "7855c5ce-df38-4467-95df-146d3f9a71ce", 00:14:17.609 "assigned_rate_limits": { 00:14:17.609 "rw_ios_per_sec": 0, 00:14:17.609 "rw_mbytes_per_sec": 0, 00:14:17.609 "r_mbytes_per_sec": 0, 00:14:17.609 "w_mbytes_per_sec": 0 00:14:17.609 }, 00:14:17.609 "claimed": false, 00:14:17.609 "zoned": false, 00:14:17.609 "supported_io_types": { 00:14:17.609 "read": true, 00:14:17.609 "write": true, 00:14:17.609 "unmap": false, 00:14:17.609 "flush": false, 00:14:17.609 "reset": true, 00:14:17.609 "nvme_admin": false, 00:14:17.609 "nvme_io": false, 00:14:17.609 "nvme_io_md": false, 00:14:17.609 "write_zeroes": true, 00:14:17.609 "zcopy": false, 00:14:17.609 "get_zone_info": false, 00:14:17.609 "zone_management": false, 00:14:17.609 "zone_append": false, 00:14:17.609 "compare": false, 00:14:17.609 "compare_and_write": false, 00:14:17.609 "abort": false, 00:14:17.609 "seek_hole": false, 00:14:17.609 "seek_data": false, 00:14:17.609 "copy": false, 00:14:17.609 "nvme_iov_md": false 00:14:17.609 }, 00:14:17.609 "driver_specific": { 00:14:17.609 "raid": { 00:14:17.609 "uuid": "7855c5ce-df38-4467-95df-146d3f9a71ce", 00:14:17.609 "strip_size_kb": 64, 00:14:17.609 "state": "online", 00:14:17.609 "raid_level": "raid5f", 00:14:17.609 "superblock": true, 00:14:17.609 "num_base_bdevs": 4, 00:14:17.609 "num_base_bdevs_discovered": 4, 00:14:17.609 "num_base_bdevs_operational": 4, 00:14:17.609 "base_bdevs_list": [ 00:14:17.609 { 00:14:17.609 "name": "NewBaseBdev", 00:14:17.609 "uuid": "a855bcc3-e4e7-4e31-8516-df34dcae7610", 00:14:17.609 "is_configured": true, 00:14:17.609 "data_offset": 2048, 00:14:17.609 "data_size": 63488 00:14:17.609 }, 00:14:17.609 { 00:14:17.609 "name": "BaseBdev2", 00:14:17.609 "uuid": "afc5c893-7a8a-48e2-b1b9-af155532c0eb", 00:14:17.609 "is_configured": true, 00:14:17.609 "data_offset": 2048, 00:14:17.609 "data_size": 63488 00:14:17.609 }, 00:14:17.609 { 00:14:17.609 "name": "BaseBdev3", 00:14:17.609 "uuid": "154206a3-5647-4582-b9e9-980956cd820e", 00:14:17.609 "is_configured": true, 00:14:17.609 "data_offset": 2048, 00:14:17.609 "data_size": 63488 00:14:17.609 }, 00:14:17.609 { 00:14:17.609 "name": "BaseBdev4", 00:14:17.609 "uuid": "9dd49a41-461f-4807-847e-65f3edfe320d", 00:14:17.609 "is_configured": true, 00:14:17.609 "data_offset": 2048, 00:14:17.609 "data_size": 63488 00:14:17.609 } 00:14:17.609 ] 00:14:17.609 } 00:14:17.609 } 00:14:17.609 }' 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:17.609 BaseBdev2 00:14:17.609 BaseBdev3 00:14:17.609 BaseBdev4' 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.609 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.868 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:17.868 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:17.868 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:17.868 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:17.868 05:41:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:17.868 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.868 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.868 05:41:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.868 [2024-12-07 05:41:51.058205] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:17.868 [2024-12-07 05:41:51.058275] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:17.868 [2024-12-07 05:41:51.058352] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:17.868 [2024-12-07 05:41:51.058606] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:17.868 [2024-12-07 05:41:51.058617] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 93490 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 93490 ']' 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 93490 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 93490 00:14:17.868 killing process with pid 93490 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 93490' 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 93490 00:14:17.868 [2024-12-07 05:41:51.108865] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:17.868 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 93490 00:14:17.868 [2024-12-07 05:41:51.149329] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:18.127 05:41:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:14:18.127 ************************************ 00:14:18.127 END TEST raid5f_state_function_test_sb 00:14:18.127 ************************************ 00:14:18.127 00:14:18.127 real 0m9.358s 00:14:18.127 user 0m16.018s 00:14:18.127 sys 0m1.980s 00:14:18.127 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:18.127 05:41:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.127 05:41:51 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:14:18.127 05:41:51 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:14:18.127 05:41:51 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:18.127 05:41:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:18.127 ************************************ 00:14:18.127 START TEST raid5f_superblock_test 00:14:18.127 ************************************ 00:14:18.127 05:41:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 4 00:14:18.127 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:14:18.127 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:14:18.127 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:14:18.127 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:14:18.127 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:14:18.127 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:14:18.127 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:14:18.127 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=94138 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 94138 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 94138 ']' 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:18.128 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:18.128 05:41:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.387 [2024-12-07 05:41:51.531771] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:14:18.387 [2024-12-07 05:41:51.531961] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94138 ] 00:14:18.387 [2024-12-07 05:41:51.667158] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:18.387 [2024-12-07 05:41:51.691542] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:18.387 [2024-12-07 05:41:51.733552] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:18.387 [2024-12-07 05:41:51.733694] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.325 malloc1 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.325 [2024-12-07 05:41:52.360762] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:19.325 [2024-12-07 05:41:52.360825] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:19.325 [2024-12-07 05:41:52.360869] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:19.325 [2024-12-07 05:41:52.360889] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:19.325 [2024-12-07 05:41:52.362967] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:19.325 [2024-12-07 05:41:52.363084] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:19.325 pt1 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.325 malloc2 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.325 [2024-12-07 05:41:52.389207] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:19.325 [2024-12-07 05:41:52.389267] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:19.325 [2024-12-07 05:41:52.389286] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:19.325 [2024-12-07 05:41:52.389296] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:19.325 [2024-12-07 05:41:52.391335] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:19.325 [2024-12-07 05:41:52.391377] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:19.325 pt2 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.325 malloc3 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.325 [2024-12-07 05:41:52.417539] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:19.325 [2024-12-07 05:41:52.417689] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:19.325 [2024-12-07 05:41:52.417731] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:19.325 [2024-12-07 05:41:52.417767] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:19.325 [2024-12-07 05:41:52.419783] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:19.325 [2024-12-07 05:41:52.419858] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:19.325 pt3 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.325 malloc4 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.325 [2024-12-07 05:41:52.458052] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:19.325 [2024-12-07 05:41:52.458149] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:19.325 [2024-12-07 05:41:52.458183] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:19.325 [2024-12-07 05:41:52.458214] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:19.325 [2024-12-07 05:41:52.460355] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:19.325 [2024-12-07 05:41:52.460427] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:19.325 pt4 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.325 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.325 [2024-12-07 05:41:52.470061] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:19.325 [2024-12-07 05:41:52.471939] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:19.325 [2024-12-07 05:41:52.472007] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:19.325 [2024-12-07 05:41:52.472075] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:19.325 [2024-12-07 05:41:52.472237] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:19.325 [2024-12-07 05:41:52.472249] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:19.325 [2024-12-07 05:41:52.472482] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:19.325 [2024-12-07 05:41:52.472961] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:19.325 [2024-12-07 05:41:52.473022] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:19.325 [2024-12-07 05:41:52.473164] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:19.326 "name": "raid_bdev1", 00:14:19.326 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:19.326 "strip_size_kb": 64, 00:14:19.326 "state": "online", 00:14:19.326 "raid_level": "raid5f", 00:14:19.326 "superblock": true, 00:14:19.326 "num_base_bdevs": 4, 00:14:19.326 "num_base_bdevs_discovered": 4, 00:14:19.326 "num_base_bdevs_operational": 4, 00:14:19.326 "base_bdevs_list": [ 00:14:19.326 { 00:14:19.326 "name": "pt1", 00:14:19.326 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:19.326 "is_configured": true, 00:14:19.326 "data_offset": 2048, 00:14:19.326 "data_size": 63488 00:14:19.326 }, 00:14:19.326 { 00:14:19.326 "name": "pt2", 00:14:19.326 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:19.326 "is_configured": true, 00:14:19.326 "data_offset": 2048, 00:14:19.326 "data_size": 63488 00:14:19.326 }, 00:14:19.326 { 00:14:19.326 "name": "pt3", 00:14:19.326 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:19.326 "is_configured": true, 00:14:19.326 "data_offset": 2048, 00:14:19.326 "data_size": 63488 00:14:19.326 }, 00:14:19.326 { 00:14:19.326 "name": "pt4", 00:14:19.326 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:19.326 "is_configured": true, 00:14:19.326 "data_offset": 2048, 00:14:19.326 "data_size": 63488 00:14:19.326 } 00:14:19.326 ] 00:14:19.326 }' 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:19.326 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.584 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:14:19.584 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:19.584 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:19.584 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:19.584 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:19.584 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:19.584 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:19.584 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:19.584 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.584 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.584 [2024-12-07 05:41:52.934313] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:19.843 05:41:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.843 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:19.843 "name": "raid_bdev1", 00:14:19.843 "aliases": [ 00:14:19.843 "57bba3c5-cf66-44d4-a8c8-627a1e9db788" 00:14:19.843 ], 00:14:19.843 "product_name": "Raid Volume", 00:14:19.843 "block_size": 512, 00:14:19.843 "num_blocks": 190464, 00:14:19.843 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:19.843 "assigned_rate_limits": { 00:14:19.843 "rw_ios_per_sec": 0, 00:14:19.843 "rw_mbytes_per_sec": 0, 00:14:19.843 "r_mbytes_per_sec": 0, 00:14:19.844 "w_mbytes_per_sec": 0 00:14:19.844 }, 00:14:19.844 "claimed": false, 00:14:19.844 "zoned": false, 00:14:19.844 "supported_io_types": { 00:14:19.844 "read": true, 00:14:19.844 "write": true, 00:14:19.844 "unmap": false, 00:14:19.844 "flush": false, 00:14:19.844 "reset": true, 00:14:19.844 "nvme_admin": false, 00:14:19.844 "nvme_io": false, 00:14:19.844 "nvme_io_md": false, 00:14:19.844 "write_zeroes": true, 00:14:19.844 "zcopy": false, 00:14:19.844 "get_zone_info": false, 00:14:19.844 "zone_management": false, 00:14:19.844 "zone_append": false, 00:14:19.844 "compare": false, 00:14:19.844 "compare_and_write": false, 00:14:19.844 "abort": false, 00:14:19.844 "seek_hole": false, 00:14:19.844 "seek_data": false, 00:14:19.844 "copy": false, 00:14:19.844 "nvme_iov_md": false 00:14:19.844 }, 00:14:19.844 "driver_specific": { 00:14:19.844 "raid": { 00:14:19.844 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:19.844 "strip_size_kb": 64, 00:14:19.844 "state": "online", 00:14:19.844 "raid_level": "raid5f", 00:14:19.844 "superblock": true, 00:14:19.844 "num_base_bdevs": 4, 00:14:19.844 "num_base_bdevs_discovered": 4, 00:14:19.844 "num_base_bdevs_operational": 4, 00:14:19.844 "base_bdevs_list": [ 00:14:19.844 { 00:14:19.844 "name": "pt1", 00:14:19.844 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:19.844 "is_configured": true, 00:14:19.844 "data_offset": 2048, 00:14:19.844 "data_size": 63488 00:14:19.844 }, 00:14:19.844 { 00:14:19.844 "name": "pt2", 00:14:19.844 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:19.844 "is_configured": true, 00:14:19.844 "data_offset": 2048, 00:14:19.844 "data_size": 63488 00:14:19.844 }, 00:14:19.844 { 00:14:19.844 "name": "pt3", 00:14:19.844 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:19.844 "is_configured": true, 00:14:19.844 "data_offset": 2048, 00:14:19.844 "data_size": 63488 00:14:19.844 }, 00:14:19.844 { 00:14:19.844 "name": "pt4", 00:14:19.844 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:19.844 "is_configured": true, 00:14:19.844 "data_offset": 2048, 00:14:19.844 "data_size": 63488 00:14:19.844 } 00:14:19.844 ] 00:14:19.844 } 00:14:19.844 } 00:14:19.844 }' 00:14:19.844 05:41:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:19.844 pt2 00:14:19.844 pt3 00:14:19.844 pt4' 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.844 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:14:20.104 [2024-12-07 05:41:53.237750] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=57bba3c5-cf66-44d4-a8c8-627a1e9db788 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 57bba3c5-cf66-44d4-a8c8-627a1e9db788 ']' 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.104 [2024-12-07 05:41:53.273510] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:20.104 [2024-12-07 05:41:53.273583] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:20.104 [2024-12-07 05:41:53.273670] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:20.104 [2024-12-07 05:41:53.273758] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:20.104 [2024-12-07 05:41:53.273768] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:14:20.104 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.105 [2024-12-07 05:41:53.441255] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:14:20.105 [2024-12-07 05:41:53.443109] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:14:20.105 [2024-12-07 05:41:53.443203] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:14:20.105 [2024-12-07 05:41:53.443249] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:14:20.105 [2024-12-07 05:41:53.443319] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:14:20.105 [2024-12-07 05:41:53.443393] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:14:20.105 [2024-12-07 05:41:53.443430] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:14:20.105 [2024-12-07 05:41:53.443446] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:14:20.105 [2024-12-07 05:41:53.443459] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:20.105 [2024-12-07 05:41:53.443471] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:14:20.105 request: 00:14:20.105 { 00:14:20.105 "name": "raid_bdev1", 00:14:20.105 "raid_level": "raid5f", 00:14:20.105 "base_bdevs": [ 00:14:20.105 "malloc1", 00:14:20.105 "malloc2", 00:14:20.105 "malloc3", 00:14:20.105 "malloc4" 00:14:20.105 ], 00:14:20.105 "strip_size_kb": 64, 00:14:20.105 "superblock": false, 00:14:20.105 "method": "bdev_raid_create", 00:14:20.105 "req_id": 1 00:14:20.105 } 00:14:20.105 Got JSON-RPC error response 00:14:20.105 response: 00:14:20.105 { 00:14:20.105 "code": -17, 00:14:20.105 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:14:20.105 } 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:14:20.105 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.366 [2024-12-07 05:41:53.509109] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:20.366 [2024-12-07 05:41:53.509217] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:20.366 [2024-12-07 05:41:53.509256] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:20.366 [2024-12-07 05:41:53.509283] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:20.366 [2024-12-07 05:41:53.511357] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:20.366 [2024-12-07 05:41:53.511441] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:20.366 [2024-12-07 05:41:53.511524] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:20.366 [2024-12-07 05:41:53.511584] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:20.366 pt1 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:20.366 "name": "raid_bdev1", 00:14:20.366 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:20.366 "strip_size_kb": 64, 00:14:20.366 "state": "configuring", 00:14:20.366 "raid_level": "raid5f", 00:14:20.366 "superblock": true, 00:14:20.366 "num_base_bdevs": 4, 00:14:20.366 "num_base_bdevs_discovered": 1, 00:14:20.366 "num_base_bdevs_operational": 4, 00:14:20.366 "base_bdevs_list": [ 00:14:20.366 { 00:14:20.366 "name": "pt1", 00:14:20.366 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:20.366 "is_configured": true, 00:14:20.366 "data_offset": 2048, 00:14:20.366 "data_size": 63488 00:14:20.366 }, 00:14:20.366 { 00:14:20.366 "name": null, 00:14:20.366 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:20.366 "is_configured": false, 00:14:20.366 "data_offset": 2048, 00:14:20.366 "data_size": 63488 00:14:20.366 }, 00:14:20.366 { 00:14:20.366 "name": null, 00:14:20.366 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:20.366 "is_configured": false, 00:14:20.366 "data_offset": 2048, 00:14:20.366 "data_size": 63488 00:14:20.366 }, 00:14:20.366 { 00:14:20.366 "name": null, 00:14:20.366 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:20.366 "is_configured": false, 00:14:20.366 "data_offset": 2048, 00:14:20.366 "data_size": 63488 00:14:20.366 } 00:14:20.366 ] 00:14:20.366 }' 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:20.366 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.626 [2024-12-07 05:41:53.960400] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:20.626 [2024-12-07 05:41:53.960502] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:20.626 [2024-12-07 05:41:53.960541] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:20.626 [2024-12-07 05:41:53.960569] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:20.626 [2024-12-07 05:41:53.961002] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:20.626 [2024-12-07 05:41:53.961060] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:20.626 [2024-12-07 05:41:53.961158] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:20.626 [2024-12-07 05:41:53.961210] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:20.626 pt2 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.626 [2024-12-07 05:41:53.972398] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.626 05:41:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:20.885 05:41:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.885 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:20.885 "name": "raid_bdev1", 00:14:20.885 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:20.885 "strip_size_kb": 64, 00:14:20.885 "state": "configuring", 00:14:20.885 "raid_level": "raid5f", 00:14:20.885 "superblock": true, 00:14:20.885 "num_base_bdevs": 4, 00:14:20.885 "num_base_bdevs_discovered": 1, 00:14:20.885 "num_base_bdevs_operational": 4, 00:14:20.885 "base_bdevs_list": [ 00:14:20.885 { 00:14:20.885 "name": "pt1", 00:14:20.885 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:20.885 "is_configured": true, 00:14:20.885 "data_offset": 2048, 00:14:20.885 "data_size": 63488 00:14:20.885 }, 00:14:20.885 { 00:14:20.885 "name": null, 00:14:20.885 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:20.885 "is_configured": false, 00:14:20.885 "data_offset": 0, 00:14:20.885 "data_size": 63488 00:14:20.885 }, 00:14:20.885 { 00:14:20.885 "name": null, 00:14:20.885 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:20.885 "is_configured": false, 00:14:20.885 "data_offset": 2048, 00:14:20.885 "data_size": 63488 00:14:20.885 }, 00:14:20.885 { 00:14:20.885 "name": null, 00:14:20.885 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:20.885 "is_configured": false, 00:14:20.885 "data_offset": 2048, 00:14:20.885 "data_size": 63488 00:14:20.885 } 00:14:20.885 ] 00:14:20.885 }' 00:14:20.885 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:20.885 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.144 [2024-12-07 05:41:54.363722] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:21.144 [2024-12-07 05:41:54.363820] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:21.144 [2024-12-07 05:41:54.363851] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:14:21.144 [2024-12-07 05:41:54.363879] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:21.144 [2024-12-07 05:41:54.364266] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:21.144 [2024-12-07 05:41:54.364328] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:21.144 [2024-12-07 05:41:54.364415] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:21.144 [2024-12-07 05:41:54.364450] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:21.144 pt2 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.144 [2024-12-07 05:41:54.375673] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:21.144 [2024-12-07 05:41:54.375769] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:21.144 [2024-12-07 05:41:54.375798] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:14:21.144 [2024-12-07 05:41:54.375826] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:21.144 [2024-12-07 05:41:54.376158] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:21.144 [2024-12-07 05:41:54.376217] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:21.144 [2024-12-07 05:41:54.376294] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:21.144 [2024-12-07 05:41:54.376345] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:21.144 pt3 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.144 [2024-12-07 05:41:54.387673] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:21.144 [2024-12-07 05:41:54.387772] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:21.144 [2024-12-07 05:41:54.387802] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:14:21.144 [2024-12-07 05:41:54.387829] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:21.144 [2024-12-07 05:41:54.388126] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:21.144 [2024-12-07 05:41:54.388184] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:21.144 [2024-12-07 05:41:54.388261] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:21.144 [2024-12-07 05:41:54.388309] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:21.144 [2024-12-07 05:41:54.388430] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:21.144 [2024-12-07 05:41:54.388471] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:21.144 [2024-12-07 05:41:54.388727] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:21.144 [2024-12-07 05:41:54.389231] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:21.144 [2024-12-07 05:41:54.389281] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:14:21.144 [2024-12-07 05:41:54.389415] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:21.144 pt4 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:21.144 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:21.145 "name": "raid_bdev1", 00:14:21.145 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:21.145 "strip_size_kb": 64, 00:14:21.145 "state": "online", 00:14:21.145 "raid_level": "raid5f", 00:14:21.145 "superblock": true, 00:14:21.145 "num_base_bdevs": 4, 00:14:21.145 "num_base_bdevs_discovered": 4, 00:14:21.145 "num_base_bdevs_operational": 4, 00:14:21.145 "base_bdevs_list": [ 00:14:21.145 { 00:14:21.145 "name": "pt1", 00:14:21.145 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:21.145 "is_configured": true, 00:14:21.145 "data_offset": 2048, 00:14:21.145 "data_size": 63488 00:14:21.145 }, 00:14:21.145 { 00:14:21.145 "name": "pt2", 00:14:21.145 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:21.145 "is_configured": true, 00:14:21.145 "data_offset": 2048, 00:14:21.145 "data_size": 63488 00:14:21.145 }, 00:14:21.145 { 00:14:21.145 "name": "pt3", 00:14:21.145 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:21.145 "is_configured": true, 00:14:21.145 "data_offset": 2048, 00:14:21.145 "data_size": 63488 00:14:21.145 }, 00:14:21.145 { 00:14:21.145 "name": "pt4", 00:14:21.145 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:21.145 "is_configured": true, 00:14:21.145 "data_offset": 2048, 00:14:21.145 "data_size": 63488 00:14:21.145 } 00:14:21.145 ] 00:14:21.145 }' 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:21.145 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.711 [2024-12-07 05:41:54.855053] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.711 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:21.711 "name": "raid_bdev1", 00:14:21.711 "aliases": [ 00:14:21.711 "57bba3c5-cf66-44d4-a8c8-627a1e9db788" 00:14:21.711 ], 00:14:21.711 "product_name": "Raid Volume", 00:14:21.711 "block_size": 512, 00:14:21.711 "num_blocks": 190464, 00:14:21.711 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:21.711 "assigned_rate_limits": { 00:14:21.711 "rw_ios_per_sec": 0, 00:14:21.711 "rw_mbytes_per_sec": 0, 00:14:21.711 "r_mbytes_per_sec": 0, 00:14:21.711 "w_mbytes_per_sec": 0 00:14:21.711 }, 00:14:21.711 "claimed": false, 00:14:21.711 "zoned": false, 00:14:21.711 "supported_io_types": { 00:14:21.711 "read": true, 00:14:21.711 "write": true, 00:14:21.711 "unmap": false, 00:14:21.711 "flush": false, 00:14:21.712 "reset": true, 00:14:21.712 "nvme_admin": false, 00:14:21.712 "nvme_io": false, 00:14:21.712 "nvme_io_md": false, 00:14:21.712 "write_zeroes": true, 00:14:21.712 "zcopy": false, 00:14:21.712 "get_zone_info": false, 00:14:21.712 "zone_management": false, 00:14:21.712 "zone_append": false, 00:14:21.712 "compare": false, 00:14:21.712 "compare_and_write": false, 00:14:21.712 "abort": false, 00:14:21.712 "seek_hole": false, 00:14:21.712 "seek_data": false, 00:14:21.712 "copy": false, 00:14:21.712 "nvme_iov_md": false 00:14:21.712 }, 00:14:21.712 "driver_specific": { 00:14:21.712 "raid": { 00:14:21.712 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:21.712 "strip_size_kb": 64, 00:14:21.712 "state": "online", 00:14:21.712 "raid_level": "raid5f", 00:14:21.712 "superblock": true, 00:14:21.712 "num_base_bdevs": 4, 00:14:21.712 "num_base_bdevs_discovered": 4, 00:14:21.712 "num_base_bdevs_operational": 4, 00:14:21.712 "base_bdevs_list": [ 00:14:21.712 { 00:14:21.712 "name": "pt1", 00:14:21.712 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:21.712 "is_configured": true, 00:14:21.712 "data_offset": 2048, 00:14:21.712 "data_size": 63488 00:14:21.712 }, 00:14:21.712 { 00:14:21.712 "name": "pt2", 00:14:21.712 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:21.712 "is_configured": true, 00:14:21.712 "data_offset": 2048, 00:14:21.712 "data_size": 63488 00:14:21.712 }, 00:14:21.712 { 00:14:21.712 "name": "pt3", 00:14:21.712 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:21.712 "is_configured": true, 00:14:21.712 "data_offset": 2048, 00:14:21.712 "data_size": 63488 00:14:21.712 }, 00:14:21.712 { 00:14:21.712 "name": "pt4", 00:14:21.712 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:21.712 "is_configured": true, 00:14:21.712 "data_offset": 2048, 00:14:21.712 "data_size": 63488 00:14:21.712 } 00:14:21.712 ] 00:14:21.712 } 00:14:21.712 } 00:14:21.712 }' 00:14:21.712 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:21.712 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:21.712 pt2 00:14:21.712 pt3 00:14:21.712 pt4' 00:14:21.712 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:21.712 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:21.712 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:21.712 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:21.712 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.712 05:41:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.712 05:41:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:21.712 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.712 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:21.712 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:21.712 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:21.712 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:21.712 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.712 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:21.712 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.971 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.971 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:21.971 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:21.971 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:21.971 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:21.971 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.971 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.971 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:21.971 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.971 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:21.971 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.972 [2024-12-07 05:41:55.210599] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 57bba3c5-cf66-44d4-a8c8-627a1e9db788 '!=' 57bba3c5-cf66-44d4-a8c8-627a1e9db788 ']' 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.972 [2024-12-07 05:41:55.254395] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:21.972 "name": "raid_bdev1", 00:14:21.972 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:21.972 "strip_size_kb": 64, 00:14:21.972 "state": "online", 00:14:21.972 "raid_level": "raid5f", 00:14:21.972 "superblock": true, 00:14:21.972 "num_base_bdevs": 4, 00:14:21.972 "num_base_bdevs_discovered": 3, 00:14:21.972 "num_base_bdevs_operational": 3, 00:14:21.972 "base_bdevs_list": [ 00:14:21.972 { 00:14:21.972 "name": null, 00:14:21.972 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.972 "is_configured": false, 00:14:21.972 "data_offset": 0, 00:14:21.972 "data_size": 63488 00:14:21.972 }, 00:14:21.972 { 00:14:21.972 "name": "pt2", 00:14:21.972 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:21.972 "is_configured": true, 00:14:21.972 "data_offset": 2048, 00:14:21.972 "data_size": 63488 00:14:21.972 }, 00:14:21.972 { 00:14:21.972 "name": "pt3", 00:14:21.972 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:21.972 "is_configured": true, 00:14:21.972 "data_offset": 2048, 00:14:21.972 "data_size": 63488 00:14:21.972 }, 00:14:21.972 { 00:14:21.972 "name": "pt4", 00:14:21.972 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:21.972 "is_configured": true, 00:14:21.972 "data_offset": 2048, 00:14:21.972 "data_size": 63488 00:14:21.972 } 00:14:21.972 ] 00:14:21.972 }' 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:21.972 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.540 [2024-12-07 05:41:55.677614] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:22.540 [2024-12-07 05:41:55.677649] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:22.540 [2024-12-07 05:41:55.677715] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:22.540 [2024-12-07 05:41:55.677780] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:22.540 [2024-12-07 05:41:55.677792] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.540 [2024-12-07 05:41:55.777440] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:22.540 [2024-12-07 05:41:55.777536] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:22.540 [2024-12-07 05:41:55.777556] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:14:22.540 [2024-12-07 05:41:55.777566] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:22.540 [2024-12-07 05:41:55.779756] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:22.540 [2024-12-07 05:41:55.779794] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:22.540 [2024-12-07 05:41:55.779860] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:22.540 [2024-12-07 05:41:55.779895] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:22.540 pt2 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:22.540 "name": "raid_bdev1", 00:14:22.540 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:22.540 "strip_size_kb": 64, 00:14:22.540 "state": "configuring", 00:14:22.540 "raid_level": "raid5f", 00:14:22.540 "superblock": true, 00:14:22.540 "num_base_bdevs": 4, 00:14:22.540 "num_base_bdevs_discovered": 1, 00:14:22.540 "num_base_bdevs_operational": 3, 00:14:22.540 "base_bdevs_list": [ 00:14:22.540 { 00:14:22.540 "name": null, 00:14:22.540 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.540 "is_configured": false, 00:14:22.540 "data_offset": 2048, 00:14:22.540 "data_size": 63488 00:14:22.540 }, 00:14:22.540 { 00:14:22.540 "name": "pt2", 00:14:22.540 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:22.540 "is_configured": true, 00:14:22.540 "data_offset": 2048, 00:14:22.540 "data_size": 63488 00:14:22.540 }, 00:14:22.540 { 00:14:22.540 "name": null, 00:14:22.540 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:22.540 "is_configured": false, 00:14:22.540 "data_offset": 2048, 00:14:22.540 "data_size": 63488 00:14:22.540 }, 00:14:22.540 { 00:14:22.540 "name": null, 00:14:22.540 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:22.540 "is_configured": false, 00:14:22.540 "data_offset": 2048, 00:14:22.540 "data_size": 63488 00:14:22.540 } 00:14:22.540 ] 00:14:22.540 }' 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:22.540 05:41:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.108 [2024-12-07 05:41:56.208751] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:23.108 [2024-12-07 05:41:56.208871] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:23.108 [2024-12-07 05:41:56.208906] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:14:23.108 [2024-12-07 05:41:56.208937] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:23.108 [2024-12-07 05:41:56.209318] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:23.108 [2024-12-07 05:41:56.209378] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:23.108 [2024-12-07 05:41:56.209474] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:23.108 [2024-12-07 05:41:56.209527] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:23.108 pt3 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.108 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:23.109 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.109 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.109 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:23.109 "name": "raid_bdev1", 00:14:23.109 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:23.109 "strip_size_kb": 64, 00:14:23.109 "state": "configuring", 00:14:23.109 "raid_level": "raid5f", 00:14:23.109 "superblock": true, 00:14:23.109 "num_base_bdevs": 4, 00:14:23.109 "num_base_bdevs_discovered": 2, 00:14:23.109 "num_base_bdevs_operational": 3, 00:14:23.109 "base_bdevs_list": [ 00:14:23.109 { 00:14:23.109 "name": null, 00:14:23.109 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.109 "is_configured": false, 00:14:23.109 "data_offset": 2048, 00:14:23.109 "data_size": 63488 00:14:23.109 }, 00:14:23.109 { 00:14:23.109 "name": "pt2", 00:14:23.109 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:23.109 "is_configured": true, 00:14:23.109 "data_offset": 2048, 00:14:23.109 "data_size": 63488 00:14:23.109 }, 00:14:23.109 { 00:14:23.109 "name": "pt3", 00:14:23.109 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:23.109 "is_configured": true, 00:14:23.109 "data_offset": 2048, 00:14:23.109 "data_size": 63488 00:14:23.109 }, 00:14:23.109 { 00:14:23.109 "name": null, 00:14:23.109 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:23.109 "is_configured": false, 00:14:23.109 "data_offset": 2048, 00:14:23.109 "data_size": 63488 00:14:23.109 } 00:14:23.109 ] 00:14:23.109 }' 00:14:23.109 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:23.109 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.369 [2024-12-07 05:41:56.628042] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:23.369 [2024-12-07 05:41:56.628143] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:23.369 [2024-12-07 05:41:56.628179] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:14:23.369 [2024-12-07 05:41:56.628208] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:23.369 [2024-12-07 05:41:56.628580] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:23.369 [2024-12-07 05:41:56.628653] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:23.369 [2024-12-07 05:41:56.628749] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:23.369 [2024-12-07 05:41:56.628801] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:23.369 [2024-12-07 05:41:56.628913] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:23.369 [2024-12-07 05:41:56.628953] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:23.369 [2024-12-07 05:41:56.629193] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:14:23.369 [2024-12-07 05:41:56.629773] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:23.369 [2024-12-07 05:41:56.629825] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:14:23.369 [2024-12-07 05:41:56.630082] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:23.369 pt4 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:23.369 "name": "raid_bdev1", 00:14:23.369 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:23.369 "strip_size_kb": 64, 00:14:23.369 "state": "online", 00:14:23.369 "raid_level": "raid5f", 00:14:23.369 "superblock": true, 00:14:23.369 "num_base_bdevs": 4, 00:14:23.369 "num_base_bdevs_discovered": 3, 00:14:23.369 "num_base_bdevs_operational": 3, 00:14:23.369 "base_bdevs_list": [ 00:14:23.369 { 00:14:23.369 "name": null, 00:14:23.369 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.369 "is_configured": false, 00:14:23.369 "data_offset": 2048, 00:14:23.369 "data_size": 63488 00:14:23.369 }, 00:14:23.369 { 00:14:23.369 "name": "pt2", 00:14:23.369 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:23.369 "is_configured": true, 00:14:23.369 "data_offset": 2048, 00:14:23.369 "data_size": 63488 00:14:23.369 }, 00:14:23.369 { 00:14:23.369 "name": "pt3", 00:14:23.369 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:23.369 "is_configured": true, 00:14:23.369 "data_offset": 2048, 00:14:23.369 "data_size": 63488 00:14:23.369 }, 00:14:23.369 { 00:14:23.369 "name": "pt4", 00:14:23.369 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:23.369 "is_configured": true, 00:14:23.369 "data_offset": 2048, 00:14:23.369 "data_size": 63488 00:14:23.369 } 00:14:23.369 ] 00:14:23.369 }' 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:23.369 05:41:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.939 [2024-12-07 05:41:57.039367] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:23.939 [2024-12-07 05:41:57.039396] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:23.939 [2024-12-07 05:41:57.039467] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:23.939 [2024-12-07 05:41:57.039541] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:23.939 [2024-12-07 05:41:57.039550] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.939 [2024-12-07 05:41:57.111250] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:23.939 [2024-12-07 05:41:57.111323] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:23.939 [2024-12-07 05:41:57.111344] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:14:23.939 [2024-12-07 05:41:57.111354] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:23.939 [2024-12-07 05:41:57.113516] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:23.939 [2024-12-07 05:41:57.113607] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:23.939 [2024-12-07 05:41:57.113697] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:23.939 [2024-12-07 05:41:57.113733] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:23.939 [2024-12-07 05:41:57.113835] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:14:23.939 [2024-12-07 05:41:57.113847] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:23.939 [2024-12-07 05:41:57.113865] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:14:23.939 [2024-12-07 05:41:57.113905] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:23.939 [2024-12-07 05:41:57.114002] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:23.939 pt1 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.939 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:23.939 "name": "raid_bdev1", 00:14:23.939 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:23.939 "strip_size_kb": 64, 00:14:23.939 "state": "configuring", 00:14:23.939 "raid_level": "raid5f", 00:14:23.939 "superblock": true, 00:14:23.939 "num_base_bdevs": 4, 00:14:23.939 "num_base_bdevs_discovered": 2, 00:14:23.939 "num_base_bdevs_operational": 3, 00:14:23.939 "base_bdevs_list": [ 00:14:23.939 { 00:14:23.939 "name": null, 00:14:23.939 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.939 "is_configured": false, 00:14:23.939 "data_offset": 2048, 00:14:23.939 "data_size": 63488 00:14:23.939 }, 00:14:23.939 { 00:14:23.939 "name": "pt2", 00:14:23.939 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:23.939 "is_configured": true, 00:14:23.939 "data_offset": 2048, 00:14:23.940 "data_size": 63488 00:14:23.940 }, 00:14:23.940 { 00:14:23.940 "name": "pt3", 00:14:23.940 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:23.940 "is_configured": true, 00:14:23.940 "data_offset": 2048, 00:14:23.940 "data_size": 63488 00:14:23.940 }, 00:14:23.940 { 00:14:23.940 "name": null, 00:14:23.940 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:23.940 "is_configured": false, 00:14:23.940 "data_offset": 2048, 00:14:23.940 "data_size": 63488 00:14:23.940 } 00:14:23.940 ] 00:14:23.940 }' 00:14:23.940 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:23.940 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.199 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:14:24.199 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:24.199 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.199 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.199 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.459 [2024-12-07 05:41:57.582486] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:24.459 [2024-12-07 05:41:57.582592] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:24.459 [2024-12-07 05:41:57.582643] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:14:24.459 [2024-12-07 05:41:57.582676] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:24.459 [2024-12-07 05:41:57.583097] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:24.459 [2024-12-07 05:41:57.583165] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:24.459 [2024-12-07 05:41:57.583265] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:24.459 [2024-12-07 05:41:57.583328] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:24.459 [2024-12-07 05:41:57.583458] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:14:24.459 [2024-12-07 05:41:57.583506] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:24.459 [2024-12-07 05:41:57.583805] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:14:24.459 [2024-12-07 05:41:57.584457] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:14:24.459 [2024-12-07 05:41:57.584525] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:14:24.459 [2024-12-07 05:41:57.584774] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:24.459 pt4 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:24.459 "name": "raid_bdev1", 00:14:24.459 "uuid": "57bba3c5-cf66-44d4-a8c8-627a1e9db788", 00:14:24.459 "strip_size_kb": 64, 00:14:24.459 "state": "online", 00:14:24.459 "raid_level": "raid5f", 00:14:24.459 "superblock": true, 00:14:24.459 "num_base_bdevs": 4, 00:14:24.459 "num_base_bdevs_discovered": 3, 00:14:24.459 "num_base_bdevs_operational": 3, 00:14:24.459 "base_bdevs_list": [ 00:14:24.459 { 00:14:24.459 "name": null, 00:14:24.459 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.459 "is_configured": false, 00:14:24.459 "data_offset": 2048, 00:14:24.459 "data_size": 63488 00:14:24.459 }, 00:14:24.459 { 00:14:24.459 "name": "pt2", 00:14:24.459 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:24.459 "is_configured": true, 00:14:24.459 "data_offset": 2048, 00:14:24.459 "data_size": 63488 00:14:24.459 }, 00:14:24.459 { 00:14:24.459 "name": "pt3", 00:14:24.459 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:24.459 "is_configured": true, 00:14:24.459 "data_offset": 2048, 00:14:24.459 "data_size": 63488 00:14:24.459 }, 00:14:24.459 { 00:14:24.459 "name": "pt4", 00:14:24.459 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:24.459 "is_configured": true, 00:14:24.459 "data_offset": 2048, 00:14:24.459 "data_size": 63488 00:14:24.459 } 00:14:24.459 ] 00:14:24.459 }' 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:24.459 05:41:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.719 05:41:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:14:24.719 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.719 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.719 05:41:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:24.719 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.978 [2024-12-07 05:41:58.097819] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 57bba3c5-cf66-44d4-a8c8-627a1e9db788 '!=' 57bba3c5-cf66-44d4-a8c8-627a1e9db788 ']' 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 94138 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 94138 ']' 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 94138 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 94138 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 94138' 00:14:24.978 killing process with pid 94138 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 94138 00:14:24.978 [2024-12-07 05:41:58.180825] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:24.978 [2024-12-07 05:41:58.180899] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:24.978 [2024-12-07 05:41:58.180977] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:24.978 [2024-12-07 05:41:58.180986] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:14:24.978 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 94138 00:14:24.978 [2024-12-07 05:41:58.224357] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:25.238 05:41:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:14:25.238 00:14:25.238 real 0m7.000s 00:14:25.238 user 0m11.772s 00:14:25.238 sys 0m1.506s 00:14:25.238 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:25.238 ************************************ 00:14:25.238 END TEST raid5f_superblock_test 00:14:25.238 ************************************ 00:14:25.238 05:41:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.238 05:41:58 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:14:25.238 05:41:58 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:14:25.238 05:41:58 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:25.238 05:41:58 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:25.238 05:41:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:25.238 ************************************ 00:14:25.238 START TEST raid5f_rebuild_test 00:14:25.238 ************************************ 00:14:25.238 05:41:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 false false true 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=94607 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 94607 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 94607 ']' 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:25.239 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:25.239 05:41:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.497 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:25.497 Zero copy mechanism will not be used. 00:14:25.497 [2024-12-07 05:41:58.612678] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:14:25.497 [2024-12-07 05:41:58.612800] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94607 ] 00:14:25.497 [2024-12-07 05:41:58.768244] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:25.497 [2024-12-07 05:41:58.792806] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:25.497 [2024-12-07 05:41:58.834741] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:25.497 [2024-12-07 05:41:58.834777] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:26.065 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:26.065 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:14:26.065 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:26.065 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:26.065 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.065 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.324 BaseBdev1_malloc 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.324 [2024-12-07 05:41:59.442011] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:26.324 [2024-12-07 05:41:59.442076] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:26.324 [2024-12-07 05:41:59.442124] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:26.324 [2024-12-07 05:41:59.442139] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:26.324 [2024-12-07 05:41:59.444197] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:26.324 [2024-12-07 05:41:59.444285] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:26.324 BaseBdev1 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.324 BaseBdev2_malloc 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.324 [2024-12-07 05:41:59.466428] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:26.324 [2024-12-07 05:41:59.466481] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:26.324 [2024-12-07 05:41:59.466519] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:26.324 [2024-12-07 05:41:59.466529] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:26.324 [2024-12-07 05:41:59.468590] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:26.324 [2024-12-07 05:41:59.468663] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:26.324 BaseBdev2 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.324 BaseBdev3_malloc 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.324 [2024-12-07 05:41:59.494864] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:26.324 [2024-12-07 05:41:59.494918] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:26.324 [2024-12-07 05:41:59.494955] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:26.324 [2024-12-07 05:41:59.494963] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:26.324 [2024-12-07 05:41:59.497017] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:26.324 [2024-12-07 05:41:59.497051] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:26.324 BaseBdev3 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.324 BaseBdev4_malloc 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.324 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.325 [2024-12-07 05:41:59.539491] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:26.325 [2024-12-07 05:41:59.539545] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:26.325 [2024-12-07 05:41:59.539568] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:26.325 [2024-12-07 05:41:59.539578] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:26.325 [2024-12-07 05:41:59.541851] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:26.325 [2024-12-07 05:41:59.541887] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:26.325 BaseBdev4 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.325 spare_malloc 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.325 spare_delay 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.325 [2024-12-07 05:41:59.579852] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:26.325 [2024-12-07 05:41:59.579899] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:26.325 [2024-12-07 05:41:59.579917] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:26.325 [2024-12-07 05:41:59.579926] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:26.325 [2024-12-07 05:41:59.581981] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:26.325 [2024-12-07 05:41:59.582018] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:26.325 spare 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.325 [2024-12-07 05:41:59.591909] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:26.325 [2024-12-07 05:41:59.593760] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:26.325 [2024-12-07 05:41:59.593818] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:26.325 [2024-12-07 05:41:59.593864] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:26.325 [2024-12-07 05:41:59.593950] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:26.325 [2024-12-07 05:41:59.593959] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:26.325 [2024-12-07 05:41:59.594187] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:26.325 [2024-12-07 05:41:59.594633] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:26.325 [2024-12-07 05:41:59.594647] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:26.325 [2024-12-07 05:41:59.594759] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:26.325 "name": "raid_bdev1", 00:14:26.325 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:26.325 "strip_size_kb": 64, 00:14:26.325 "state": "online", 00:14:26.325 "raid_level": "raid5f", 00:14:26.325 "superblock": false, 00:14:26.325 "num_base_bdevs": 4, 00:14:26.325 "num_base_bdevs_discovered": 4, 00:14:26.325 "num_base_bdevs_operational": 4, 00:14:26.325 "base_bdevs_list": [ 00:14:26.325 { 00:14:26.325 "name": "BaseBdev1", 00:14:26.325 "uuid": "f9d45aa0-407e-5d6c-91ff-ca77a550668e", 00:14:26.325 "is_configured": true, 00:14:26.325 "data_offset": 0, 00:14:26.325 "data_size": 65536 00:14:26.325 }, 00:14:26.325 { 00:14:26.325 "name": "BaseBdev2", 00:14:26.325 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:26.325 "is_configured": true, 00:14:26.325 "data_offset": 0, 00:14:26.325 "data_size": 65536 00:14:26.325 }, 00:14:26.325 { 00:14:26.325 "name": "BaseBdev3", 00:14:26.325 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:26.325 "is_configured": true, 00:14:26.325 "data_offset": 0, 00:14:26.325 "data_size": 65536 00:14:26.325 }, 00:14:26.325 { 00:14:26.325 "name": "BaseBdev4", 00:14:26.325 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:26.325 "is_configured": true, 00:14:26.325 "data_offset": 0, 00:14:26.325 "data_size": 65536 00:14:26.325 } 00:14:26.325 ] 00:14:26.325 }' 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:26.325 05:41:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.892 [2024-12-07 05:42:00.059883] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:26.892 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:27.181 [2024-12-07 05:42:00.323338] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:14:27.181 /dev/nbd0 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:27.181 1+0 records in 00:14:27.181 1+0 records out 00:14:27.181 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000335558 s, 12.2 MB/s 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:14:27.181 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:14:27.440 512+0 records in 00:14:27.440 512+0 records out 00:14:27.440 100663296 bytes (101 MB, 96 MiB) copied, 0.380932 s, 264 MB/s 00:14:27.440 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:27.440 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:27.440 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:27.440 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:27.440 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:27.440 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:27.440 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:27.698 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:27.698 [2024-12-07 05:42:00.979319] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:27.698 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:27.698 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:27.698 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:27.698 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:27.698 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:27.698 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:27.698 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:27.698 05:42:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:27.698 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.698 05:42:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.698 [2024-12-07 05:42:00.999362] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.698 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:27.698 "name": "raid_bdev1", 00:14:27.698 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:27.698 "strip_size_kb": 64, 00:14:27.698 "state": "online", 00:14:27.698 "raid_level": "raid5f", 00:14:27.698 "superblock": false, 00:14:27.698 "num_base_bdevs": 4, 00:14:27.698 "num_base_bdevs_discovered": 3, 00:14:27.698 "num_base_bdevs_operational": 3, 00:14:27.698 "base_bdevs_list": [ 00:14:27.698 { 00:14:27.698 "name": null, 00:14:27.698 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:27.698 "is_configured": false, 00:14:27.699 "data_offset": 0, 00:14:27.699 "data_size": 65536 00:14:27.699 }, 00:14:27.699 { 00:14:27.699 "name": "BaseBdev2", 00:14:27.699 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:27.699 "is_configured": true, 00:14:27.699 "data_offset": 0, 00:14:27.699 "data_size": 65536 00:14:27.699 }, 00:14:27.699 { 00:14:27.699 "name": "BaseBdev3", 00:14:27.699 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:27.699 "is_configured": true, 00:14:27.699 "data_offset": 0, 00:14:27.699 "data_size": 65536 00:14:27.699 }, 00:14:27.699 { 00:14:27.699 "name": "BaseBdev4", 00:14:27.699 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:27.699 "is_configured": true, 00:14:27.699 "data_offset": 0, 00:14:27.699 "data_size": 65536 00:14:27.699 } 00:14:27.699 ] 00:14:27.699 }' 00:14:27.699 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:27.699 05:42:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.265 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:28.265 05:42:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.265 05:42:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.265 [2024-12-07 05:42:01.418695] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:28.265 [2024-12-07 05:42:01.422875] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:14:28.265 05:42:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.265 05:42:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:28.265 [2024-12-07 05:42:01.425141] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:29.201 "name": "raid_bdev1", 00:14:29.201 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:29.201 "strip_size_kb": 64, 00:14:29.201 "state": "online", 00:14:29.201 "raid_level": "raid5f", 00:14:29.201 "superblock": false, 00:14:29.201 "num_base_bdevs": 4, 00:14:29.201 "num_base_bdevs_discovered": 4, 00:14:29.201 "num_base_bdevs_operational": 4, 00:14:29.201 "process": { 00:14:29.201 "type": "rebuild", 00:14:29.201 "target": "spare", 00:14:29.201 "progress": { 00:14:29.201 "blocks": 19200, 00:14:29.201 "percent": 9 00:14:29.201 } 00:14:29.201 }, 00:14:29.201 "base_bdevs_list": [ 00:14:29.201 { 00:14:29.201 "name": "spare", 00:14:29.201 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:29.201 "is_configured": true, 00:14:29.201 "data_offset": 0, 00:14:29.201 "data_size": 65536 00:14:29.201 }, 00:14:29.201 { 00:14:29.201 "name": "BaseBdev2", 00:14:29.201 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:29.201 "is_configured": true, 00:14:29.201 "data_offset": 0, 00:14:29.201 "data_size": 65536 00:14:29.201 }, 00:14:29.201 { 00:14:29.201 "name": "BaseBdev3", 00:14:29.201 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:29.201 "is_configured": true, 00:14:29.201 "data_offset": 0, 00:14:29.201 "data_size": 65536 00:14:29.201 }, 00:14:29.201 { 00:14:29.201 "name": "BaseBdev4", 00:14:29.201 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:29.201 "is_configured": true, 00:14:29.201 "data_offset": 0, 00:14:29.201 "data_size": 65536 00:14:29.201 } 00:14:29.201 ] 00:14:29.201 }' 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:29.201 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:29.460 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:29.460 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:29.460 05:42:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.460 05:42:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.460 [2024-12-07 05:42:02.586217] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:29.460 [2024-12-07 05:42:02.630994] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:29.460 [2024-12-07 05:42:02.631057] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:29.460 [2024-12-07 05:42:02.631082] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:29.460 [2024-12-07 05:42:02.631092] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:29.460 05:42:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.460 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:29.460 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:29.460 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:29.460 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:29.460 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:29.461 "name": "raid_bdev1", 00:14:29.461 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:29.461 "strip_size_kb": 64, 00:14:29.461 "state": "online", 00:14:29.461 "raid_level": "raid5f", 00:14:29.461 "superblock": false, 00:14:29.461 "num_base_bdevs": 4, 00:14:29.461 "num_base_bdevs_discovered": 3, 00:14:29.461 "num_base_bdevs_operational": 3, 00:14:29.461 "base_bdevs_list": [ 00:14:29.461 { 00:14:29.461 "name": null, 00:14:29.461 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:29.461 "is_configured": false, 00:14:29.461 "data_offset": 0, 00:14:29.461 "data_size": 65536 00:14:29.461 }, 00:14:29.461 { 00:14:29.461 "name": "BaseBdev2", 00:14:29.461 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:29.461 "is_configured": true, 00:14:29.461 "data_offset": 0, 00:14:29.461 "data_size": 65536 00:14:29.461 }, 00:14:29.461 { 00:14:29.461 "name": "BaseBdev3", 00:14:29.461 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:29.461 "is_configured": true, 00:14:29.461 "data_offset": 0, 00:14:29.461 "data_size": 65536 00:14:29.461 }, 00:14:29.461 { 00:14:29.461 "name": "BaseBdev4", 00:14:29.461 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:29.461 "is_configured": true, 00:14:29.461 "data_offset": 0, 00:14:29.461 "data_size": 65536 00:14:29.461 } 00:14:29.461 ] 00:14:29.461 }' 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:29.461 05:42:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.720 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:29.720 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:29.720 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:29.720 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:29.720 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:29.720 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.720 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.720 05:42:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.720 05:42:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.720 05:42:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.979 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:29.979 "name": "raid_bdev1", 00:14:29.979 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:29.979 "strip_size_kb": 64, 00:14:29.979 "state": "online", 00:14:29.979 "raid_level": "raid5f", 00:14:29.979 "superblock": false, 00:14:29.979 "num_base_bdevs": 4, 00:14:29.979 "num_base_bdevs_discovered": 3, 00:14:29.979 "num_base_bdevs_operational": 3, 00:14:29.979 "base_bdevs_list": [ 00:14:29.979 { 00:14:29.979 "name": null, 00:14:29.979 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:29.979 "is_configured": false, 00:14:29.979 "data_offset": 0, 00:14:29.979 "data_size": 65536 00:14:29.979 }, 00:14:29.979 { 00:14:29.979 "name": "BaseBdev2", 00:14:29.979 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:29.979 "is_configured": true, 00:14:29.979 "data_offset": 0, 00:14:29.979 "data_size": 65536 00:14:29.979 }, 00:14:29.979 { 00:14:29.979 "name": "BaseBdev3", 00:14:29.979 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:29.979 "is_configured": true, 00:14:29.979 "data_offset": 0, 00:14:29.979 "data_size": 65536 00:14:29.979 }, 00:14:29.979 { 00:14:29.979 "name": "BaseBdev4", 00:14:29.979 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:29.979 "is_configured": true, 00:14:29.979 "data_offset": 0, 00:14:29.979 "data_size": 65536 00:14:29.979 } 00:14:29.979 ] 00:14:29.979 }' 00:14:29.979 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:29.979 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:29.979 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:29.979 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:29.979 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:29.979 05:42:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.979 05:42:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.979 [2024-12-07 05:42:03.184141] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:29.979 [2024-12-07 05:42:03.187940] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027e70 00:14:29.979 [2024-12-07 05:42:03.190144] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:29.979 05:42:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.979 05:42:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:30.917 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:30.917 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:30.917 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:30.917 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:30.917 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:30.917 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.917 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:30.917 05:42:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.917 05:42:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.917 05:42:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.917 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:30.917 "name": "raid_bdev1", 00:14:30.917 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:30.917 "strip_size_kb": 64, 00:14:30.917 "state": "online", 00:14:30.917 "raid_level": "raid5f", 00:14:30.917 "superblock": false, 00:14:30.917 "num_base_bdevs": 4, 00:14:30.917 "num_base_bdevs_discovered": 4, 00:14:30.917 "num_base_bdevs_operational": 4, 00:14:30.917 "process": { 00:14:30.917 "type": "rebuild", 00:14:30.917 "target": "spare", 00:14:30.917 "progress": { 00:14:30.917 "blocks": 19200, 00:14:30.917 "percent": 9 00:14:30.917 } 00:14:30.917 }, 00:14:30.917 "base_bdevs_list": [ 00:14:30.917 { 00:14:30.917 "name": "spare", 00:14:30.917 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:30.917 "is_configured": true, 00:14:30.917 "data_offset": 0, 00:14:30.917 "data_size": 65536 00:14:30.917 }, 00:14:30.917 { 00:14:30.917 "name": "BaseBdev2", 00:14:30.917 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:30.917 "is_configured": true, 00:14:30.917 "data_offset": 0, 00:14:30.917 "data_size": 65536 00:14:30.917 }, 00:14:30.917 { 00:14:30.917 "name": "BaseBdev3", 00:14:30.917 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:30.917 "is_configured": true, 00:14:30.917 "data_offset": 0, 00:14:30.917 "data_size": 65536 00:14:30.917 }, 00:14:30.917 { 00:14:30.917 "name": "BaseBdev4", 00:14:30.917 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:30.917 "is_configured": true, 00:14:30.918 "data_offset": 0, 00:14:30.918 "data_size": 65536 00:14:30.918 } 00:14:30.918 ] 00:14:30.918 }' 00:14:30.918 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:30.918 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:30.918 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=501 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:31.177 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.178 05:42:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.178 05:42:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.178 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.178 05:42:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.178 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:31.178 "name": "raid_bdev1", 00:14:31.178 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:31.178 "strip_size_kb": 64, 00:14:31.178 "state": "online", 00:14:31.178 "raid_level": "raid5f", 00:14:31.178 "superblock": false, 00:14:31.178 "num_base_bdevs": 4, 00:14:31.178 "num_base_bdevs_discovered": 4, 00:14:31.178 "num_base_bdevs_operational": 4, 00:14:31.178 "process": { 00:14:31.178 "type": "rebuild", 00:14:31.178 "target": "spare", 00:14:31.178 "progress": { 00:14:31.178 "blocks": 21120, 00:14:31.178 "percent": 10 00:14:31.178 } 00:14:31.178 }, 00:14:31.178 "base_bdevs_list": [ 00:14:31.178 { 00:14:31.178 "name": "spare", 00:14:31.178 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:31.178 "is_configured": true, 00:14:31.178 "data_offset": 0, 00:14:31.178 "data_size": 65536 00:14:31.178 }, 00:14:31.178 { 00:14:31.178 "name": "BaseBdev2", 00:14:31.178 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:31.178 "is_configured": true, 00:14:31.178 "data_offset": 0, 00:14:31.178 "data_size": 65536 00:14:31.178 }, 00:14:31.178 { 00:14:31.178 "name": "BaseBdev3", 00:14:31.178 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:31.178 "is_configured": true, 00:14:31.178 "data_offset": 0, 00:14:31.178 "data_size": 65536 00:14:31.178 }, 00:14:31.178 { 00:14:31.178 "name": "BaseBdev4", 00:14:31.178 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:31.178 "is_configured": true, 00:14:31.178 "data_offset": 0, 00:14:31.178 "data_size": 65536 00:14:31.178 } 00:14:31.178 ] 00:14:31.178 }' 00:14:31.178 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:31.178 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:31.178 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:31.178 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:31.178 05:42:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:32.114 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:32.114 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:32.114 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:32.114 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:32.114 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:32.114 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:32.114 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:32.114 05:42:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:32.114 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:32.114 05:42:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.114 05:42:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:32.373 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:32.373 "name": "raid_bdev1", 00:14:32.373 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:32.373 "strip_size_kb": 64, 00:14:32.373 "state": "online", 00:14:32.373 "raid_level": "raid5f", 00:14:32.373 "superblock": false, 00:14:32.373 "num_base_bdevs": 4, 00:14:32.373 "num_base_bdevs_discovered": 4, 00:14:32.373 "num_base_bdevs_operational": 4, 00:14:32.373 "process": { 00:14:32.373 "type": "rebuild", 00:14:32.373 "target": "spare", 00:14:32.373 "progress": { 00:14:32.373 "blocks": 42240, 00:14:32.373 "percent": 21 00:14:32.373 } 00:14:32.373 }, 00:14:32.373 "base_bdevs_list": [ 00:14:32.373 { 00:14:32.373 "name": "spare", 00:14:32.373 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:32.373 "is_configured": true, 00:14:32.373 "data_offset": 0, 00:14:32.373 "data_size": 65536 00:14:32.373 }, 00:14:32.373 { 00:14:32.373 "name": "BaseBdev2", 00:14:32.373 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:32.373 "is_configured": true, 00:14:32.373 "data_offset": 0, 00:14:32.373 "data_size": 65536 00:14:32.373 }, 00:14:32.373 { 00:14:32.373 "name": "BaseBdev3", 00:14:32.373 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:32.373 "is_configured": true, 00:14:32.373 "data_offset": 0, 00:14:32.373 "data_size": 65536 00:14:32.373 }, 00:14:32.373 { 00:14:32.373 "name": "BaseBdev4", 00:14:32.373 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:32.373 "is_configured": true, 00:14:32.373 "data_offset": 0, 00:14:32.373 "data_size": 65536 00:14:32.373 } 00:14:32.373 ] 00:14:32.373 }' 00:14:32.373 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:32.373 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:32.373 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:32.373 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:32.373 05:42:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:33.312 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:33.313 "name": "raid_bdev1", 00:14:33.313 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:33.313 "strip_size_kb": 64, 00:14:33.313 "state": "online", 00:14:33.313 "raid_level": "raid5f", 00:14:33.313 "superblock": false, 00:14:33.313 "num_base_bdevs": 4, 00:14:33.313 "num_base_bdevs_discovered": 4, 00:14:33.313 "num_base_bdevs_operational": 4, 00:14:33.313 "process": { 00:14:33.313 "type": "rebuild", 00:14:33.313 "target": "spare", 00:14:33.313 "progress": { 00:14:33.313 "blocks": 65280, 00:14:33.313 "percent": 33 00:14:33.313 } 00:14:33.313 }, 00:14:33.313 "base_bdevs_list": [ 00:14:33.313 { 00:14:33.313 "name": "spare", 00:14:33.313 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:33.313 "is_configured": true, 00:14:33.313 "data_offset": 0, 00:14:33.313 "data_size": 65536 00:14:33.313 }, 00:14:33.313 { 00:14:33.313 "name": "BaseBdev2", 00:14:33.313 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:33.313 "is_configured": true, 00:14:33.313 "data_offset": 0, 00:14:33.313 "data_size": 65536 00:14:33.313 }, 00:14:33.313 { 00:14:33.313 "name": "BaseBdev3", 00:14:33.313 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:33.313 "is_configured": true, 00:14:33.313 "data_offset": 0, 00:14:33.313 "data_size": 65536 00:14:33.313 }, 00:14:33.313 { 00:14:33.313 "name": "BaseBdev4", 00:14:33.313 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:33.313 "is_configured": true, 00:14:33.313 "data_offset": 0, 00:14:33.313 "data_size": 65536 00:14:33.313 } 00:14:33.313 ] 00:14:33.313 }' 00:14:33.313 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:33.572 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:33.572 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:33.572 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:33.572 05:42:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.511 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:34.511 "name": "raid_bdev1", 00:14:34.511 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:34.511 "strip_size_kb": 64, 00:14:34.511 "state": "online", 00:14:34.511 "raid_level": "raid5f", 00:14:34.511 "superblock": false, 00:14:34.511 "num_base_bdevs": 4, 00:14:34.511 "num_base_bdevs_discovered": 4, 00:14:34.511 "num_base_bdevs_operational": 4, 00:14:34.511 "process": { 00:14:34.511 "type": "rebuild", 00:14:34.511 "target": "spare", 00:14:34.511 "progress": { 00:14:34.511 "blocks": 86400, 00:14:34.511 "percent": 43 00:14:34.512 } 00:14:34.512 }, 00:14:34.512 "base_bdevs_list": [ 00:14:34.512 { 00:14:34.512 "name": "spare", 00:14:34.512 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:34.512 "is_configured": true, 00:14:34.512 "data_offset": 0, 00:14:34.512 "data_size": 65536 00:14:34.512 }, 00:14:34.512 { 00:14:34.512 "name": "BaseBdev2", 00:14:34.512 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:34.512 "is_configured": true, 00:14:34.512 "data_offset": 0, 00:14:34.512 "data_size": 65536 00:14:34.512 }, 00:14:34.512 { 00:14:34.512 "name": "BaseBdev3", 00:14:34.512 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:34.512 "is_configured": true, 00:14:34.512 "data_offset": 0, 00:14:34.512 "data_size": 65536 00:14:34.512 }, 00:14:34.512 { 00:14:34.512 "name": "BaseBdev4", 00:14:34.512 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:34.512 "is_configured": true, 00:14:34.512 "data_offset": 0, 00:14:34.512 "data_size": 65536 00:14:34.512 } 00:14:34.512 ] 00:14:34.512 }' 00:14:34.512 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:34.512 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:34.512 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:34.771 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:34.771 05:42:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:35.711 "name": "raid_bdev1", 00:14:35.711 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:35.711 "strip_size_kb": 64, 00:14:35.711 "state": "online", 00:14:35.711 "raid_level": "raid5f", 00:14:35.711 "superblock": false, 00:14:35.711 "num_base_bdevs": 4, 00:14:35.711 "num_base_bdevs_discovered": 4, 00:14:35.711 "num_base_bdevs_operational": 4, 00:14:35.711 "process": { 00:14:35.711 "type": "rebuild", 00:14:35.711 "target": "spare", 00:14:35.711 "progress": { 00:14:35.711 "blocks": 107520, 00:14:35.711 "percent": 54 00:14:35.711 } 00:14:35.711 }, 00:14:35.711 "base_bdevs_list": [ 00:14:35.711 { 00:14:35.711 "name": "spare", 00:14:35.711 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:35.711 "is_configured": true, 00:14:35.711 "data_offset": 0, 00:14:35.711 "data_size": 65536 00:14:35.711 }, 00:14:35.711 { 00:14:35.711 "name": "BaseBdev2", 00:14:35.711 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:35.711 "is_configured": true, 00:14:35.711 "data_offset": 0, 00:14:35.711 "data_size": 65536 00:14:35.711 }, 00:14:35.711 { 00:14:35.711 "name": "BaseBdev3", 00:14:35.711 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:35.711 "is_configured": true, 00:14:35.711 "data_offset": 0, 00:14:35.711 "data_size": 65536 00:14:35.711 }, 00:14:35.711 { 00:14:35.711 "name": "BaseBdev4", 00:14:35.711 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:35.711 "is_configured": true, 00:14:35.711 "data_offset": 0, 00:14:35.711 "data_size": 65536 00:14:35.711 } 00:14:35.711 ] 00:14:35.711 }' 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:35.711 05:42:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:35.711 05:42:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:35.711 05:42:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:37.093 "name": "raid_bdev1", 00:14:37.093 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:37.093 "strip_size_kb": 64, 00:14:37.093 "state": "online", 00:14:37.093 "raid_level": "raid5f", 00:14:37.093 "superblock": false, 00:14:37.093 "num_base_bdevs": 4, 00:14:37.093 "num_base_bdevs_discovered": 4, 00:14:37.093 "num_base_bdevs_operational": 4, 00:14:37.093 "process": { 00:14:37.093 "type": "rebuild", 00:14:37.093 "target": "spare", 00:14:37.093 "progress": { 00:14:37.093 "blocks": 130560, 00:14:37.093 "percent": 66 00:14:37.093 } 00:14:37.093 }, 00:14:37.093 "base_bdevs_list": [ 00:14:37.093 { 00:14:37.093 "name": "spare", 00:14:37.093 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:37.093 "is_configured": true, 00:14:37.093 "data_offset": 0, 00:14:37.093 "data_size": 65536 00:14:37.093 }, 00:14:37.093 { 00:14:37.093 "name": "BaseBdev2", 00:14:37.093 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:37.093 "is_configured": true, 00:14:37.093 "data_offset": 0, 00:14:37.093 "data_size": 65536 00:14:37.093 }, 00:14:37.093 { 00:14:37.093 "name": "BaseBdev3", 00:14:37.093 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:37.093 "is_configured": true, 00:14:37.093 "data_offset": 0, 00:14:37.093 "data_size": 65536 00:14:37.093 }, 00:14:37.093 { 00:14:37.093 "name": "BaseBdev4", 00:14:37.093 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:37.093 "is_configured": true, 00:14:37.093 "data_offset": 0, 00:14:37.093 "data_size": 65536 00:14:37.093 } 00:14:37.093 ] 00:14:37.093 }' 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:37.093 05:42:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.032 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:38.032 "name": "raid_bdev1", 00:14:38.032 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:38.032 "strip_size_kb": 64, 00:14:38.032 "state": "online", 00:14:38.032 "raid_level": "raid5f", 00:14:38.032 "superblock": false, 00:14:38.032 "num_base_bdevs": 4, 00:14:38.032 "num_base_bdevs_discovered": 4, 00:14:38.032 "num_base_bdevs_operational": 4, 00:14:38.032 "process": { 00:14:38.032 "type": "rebuild", 00:14:38.032 "target": "spare", 00:14:38.032 "progress": { 00:14:38.032 "blocks": 151680, 00:14:38.032 "percent": 77 00:14:38.032 } 00:14:38.032 }, 00:14:38.032 "base_bdevs_list": [ 00:14:38.032 { 00:14:38.032 "name": "spare", 00:14:38.032 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:38.032 "is_configured": true, 00:14:38.032 "data_offset": 0, 00:14:38.032 "data_size": 65536 00:14:38.032 }, 00:14:38.032 { 00:14:38.032 "name": "BaseBdev2", 00:14:38.032 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:38.032 "is_configured": true, 00:14:38.033 "data_offset": 0, 00:14:38.033 "data_size": 65536 00:14:38.033 }, 00:14:38.033 { 00:14:38.033 "name": "BaseBdev3", 00:14:38.033 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:38.033 "is_configured": true, 00:14:38.033 "data_offset": 0, 00:14:38.033 "data_size": 65536 00:14:38.033 }, 00:14:38.033 { 00:14:38.033 "name": "BaseBdev4", 00:14:38.033 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:38.033 "is_configured": true, 00:14:38.033 "data_offset": 0, 00:14:38.033 "data_size": 65536 00:14:38.033 } 00:14:38.033 ] 00:14:38.033 }' 00:14:38.033 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:38.033 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:38.033 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:38.033 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:38.033 05:42:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.415 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:39.415 "name": "raid_bdev1", 00:14:39.415 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:39.415 "strip_size_kb": 64, 00:14:39.415 "state": "online", 00:14:39.415 "raid_level": "raid5f", 00:14:39.415 "superblock": false, 00:14:39.415 "num_base_bdevs": 4, 00:14:39.415 "num_base_bdevs_discovered": 4, 00:14:39.415 "num_base_bdevs_operational": 4, 00:14:39.415 "process": { 00:14:39.415 "type": "rebuild", 00:14:39.416 "target": "spare", 00:14:39.416 "progress": { 00:14:39.416 "blocks": 174720, 00:14:39.416 "percent": 88 00:14:39.416 } 00:14:39.416 }, 00:14:39.416 "base_bdevs_list": [ 00:14:39.416 { 00:14:39.416 "name": "spare", 00:14:39.416 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:39.416 "is_configured": true, 00:14:39.416 "data_offset": 0, 00:14:39.416 "data_size": 65536 00:14:39.416 }, 00:14:39.416 { 00:14:39.416 "name": "BaseBdev2", 00:14:39.416 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:39.416 "is_configured": true, 00:14:39.416 "data_offset": 0, 00:14:39.416 "data_size": 65536 00:14:39.416 }, 00:14:39.416 { 00:14:39.416 "name": "BaseBdev3", 00:14:39.416 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:39.416 "is_configured": true, 00:14:39.416 "data_offset": 0, 00:14:39.416 "data_size": 65536 00:14:39.416 }, 00:14:39.416 { 00:14:39.416 "name": "BaseBdev4", 00:14:39.416 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:39.416 "is_configured": true, 00:14:39.416 "data_offset": 0, 00:14:39.416 "data_size": 65536 00:14:39.416 } 00:14:39.416 ] 00:14:39.416 }' 00:14:39.416 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:39.416 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:39.416 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:39.416 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:39.416 05:42:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.355 [2024-12-07 05:42:13.533205] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:40.355 [2024-12-07 05:42:13.533278] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:40.355 [2024-12-07 05:42:13.533319] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:40.355 "name": "raid_bdev1", 00:14:40.355 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:40.355 "strip_size_kb": 64, 00:14:40.355 "state": "online", 00:14:40.355 "raid_level": "raid5f", 00:14:40.355 "superblock": false, 00:14:40.355 "num_base_bdevs": 4, 00:14:40.355 "num_base_bdevs_discovered": 4, 00:14:40.355 "num_base_bdevs_operational": 4, 00:14:40.355 "process": { 00:14:40.355 "type": "rebuild", 00:14:40.355 "target": "spare", 00:14:40.355 "progress": { 00:14:40.355 "blocks": 195840, 00:14:40.355 "percent": 99 00:14:40.355 } 00:14:40.355 }, 00:14:40.355 "base_bdevs_list": [ 00:14:40.355 { 00:14:40.355 "name": "spare", 00:14:40.355 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:40.355 "is_configured": true, 00:14:40.355 "data_offset": 0, 00:14:40.355 "data_size": 65536 00:14:40.355 }, 00:14:40.355 { 00:14:40.355 "name": "BaseBdev2", 00:14:40.355 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:40.355 "is_configured": true, 00:14:40.355 "data_offset": 0, 00:14:40.355 "data_size": 65536 00:14:40.355 }, 00:14:40.355 { 00:14:40.355 "name": "BaseBdev3", 00:14:40.355 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:40.355 "is_configured": true, 00:14:40.355 "data_offset": 0, 00:14:40.355 "data_size": 65536 00:14:40.355 }, 00:14:40.355 { 00:14:40.355 "name": "BaseBdev4", 00:14:40.355 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:40.355 "is_configured": true, 00:14:40.355 "data_offset": 0, 00:14:40.355 "data_size": 65536 00:14:40.355 } 00:14:40.355 ] 00:14:40.355 }' 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:40.355 05:42:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:41.293 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:41.293 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:41.293 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:41.293 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:41.293 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:41.293 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:41.293 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.293 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.293 05:42:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.293 05:42:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:41.554 "name": "raid_bdev1", 00:14:41.554 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:41.554 "strip_size_kb": 64, 00:14:41.554 "state": "online", 00:14:41.554 "raid_level": "raid5f", 00:14:41.554 "superblock": false, 00:14:41.554 "num_base_bdevs": 4, 00:14:41.554 "num_base_bdevs_discovered": 4, 00:14:41.554 "num_base_bdevs_operational": 4, 00:14:41.554 "base_bdevs_list": [ 00:14:41.554 { 00:14:41.554 "name": "spare", 00:14:41.554 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:41.554 "is_configured": true, 00:14:41.554 "data_offset": 0, 00:14:41.554 "data_size": 65536 00:14:41.554 }, 00:14:41.554 { 00:14:41.554 "name": "BaseBdev2", 00:14:41.554 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:41.554 "is_configured": true, 00:14:41.554 "data_offset": 0, 00:14:41.554 "data_size": 65536 00:14:41.554 }, 00:14:41.554 { 00:14:41.554 "name": "BaseBdev3", 00:14:41.554 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:41.554 "is_configured": true, 00:14:41.554 "data_offset": 0, 00:14:41.554 "data_size": 65536 00:14:41.554 }, 00:14:41.554 { 00:14:41.554 "name": "BaseBdev4", 00:14:41.554 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:41.554 "is_configured": true, 00:14:41.554 "data_offset": 0, 00:14:41.554 "data_size": 65536 00:14:41.554 } 00:14:41.554 ] 00:14:41.554 }' 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:41.554 "name": "raid_bdev1", 00:14:41.554 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:41.554 "strip_size_kb": 64, 00:14:41.554 "state": "online", 00:14:41.554 "raid_level": "raid5f", 00:14:41.554 "superblock": false, 00:14:41.554 "num_base_bdevs": 4, 00:14:41.554 "num_base_bdevs_discovered": 4, 00:14:41.554 "num_base_bdevs_operational": 4, 00:14:41.554 "base_bdevs_list": [ 00:14:41.554 { 00:14:41.554 "name": "spare", 00:14:41.554 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:41.554 "is_configured": true, 00:14:41.554 "data_offset": 0, 00:14:41.554 "data_size": 65536 00:14:41.554 }, 00:14:41.554 { 00:14:41.554 "name": "BaseBdev2", 00:14:41.554 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:41.554 "is_configured": true, 00:14:41.554 "data_offset": 0, 00:14:41.554 "data_size": 65536 00:14:41.554 }, 00:14:41.554 { 00:14:41.554 "name": "BaseBdev3", 00:14:41.554 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:41.554 "is_configured": true, 00:14:41.554 "data_offset": 0, 00:14:41.554 "data_size": 65536 00:14:41.554 }, 00:14:41.554 { 00:14:41.554 "name": "BaseBdev4", 00:14:41.554 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:41.554 "is_configured": true, 00:14:41.554 "data_offset": 0, 00:14:41.554 "data_size": 65536 00:14:41.554 } 00:14:41.554 ] 00:14:41.554 }' 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.554 05:42:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.817 05:42:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.817 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:41.817 "name": "raid_bdev1", 00:14:41.817 "uuid": "663e3254-1a3d-4ebf-b21b-c6f5f80e7a77", 00:14:41.817 "strip_size_kb": 64, 00:14:41.817 "state": "online", 00:14:41.817 "raid_level": "raid5f", 00:14:41.817 "superblock": false, 00:14:41.817 "num_base_bdevs": 4, 00:14:41.817 "num_base_bdevs_discovered": 4, 00:14:41.817 "num_base_bdevs_operational": 4, 00:14:41.817 "base_bdevs_list": [ 00:14:41.817 { 00:14:41.817 "name": "spare", 00:14:41.817 "uuid": "639d1e3e-442f-50a6-9fc0-7a5df7301d17", 00:14:41.817 "is_configured": true, 00:14:41.817 "data_offset": 0, 00:14:41.817 "data_size": 65536 00:14:41.817 }, 00:14:41.817 { 00:14:41.817 "name": "BaseBdev2", 00:14:41.817 "uuid": "2fd959a4-5d23-573e-82a2-c8176c6afe2c", 00:14:41.817 "is_configured": true, 00:14:41.817 "data_offset": 0, 00:14:41.817 "data_size": 65536 00:14:41.817 }, 00:14:41.817 { 00:14:41.817 "name": "BaseBdev3", 00:14:41.817 "uuid": "ebadc9f4-7c9d-5bc1-89ea-806e5133017b", 00:14:41.817 "is_configured": true, 00:14:41.817 "data_offset": 0, 00:14:41.817 "data_size": 65536 00:14:41.817 }, 00:14:41.817 { 00:14:41.818 "name": "BaseBdev4", 00:14:41.818 "uuid": "bb390b17-a0ea-5508-9c1a-4cc711940c17", 00:14:41.818 "is_configured": true, 00:14:41.818 "data_offset": 0, 00:14:41.818 "data_size": 65536 00:14:41.818 } 00:14:41.818 ] 00:14:41.818 }' 00:14:41.818 05:42:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:41.818 05:42:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.078 [2024-12-07 05:42:15.308606] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:42.078 [2024-12-07 05:42:15.308647] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:42.078 [2024-12-07 05:42:15.308728] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:42.078 [2024-12-07 05:42:15.308817] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:42.078 [2024-12-07 05:42:15.308847] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:42.078 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:42.338 /dev/nbd0 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:42.338 1+0 records in 00:14:42.338 1+0 records out 00:14:42.338 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000432241 s, 9.5 MB/s 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:42.338 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:42.597 /dev/nbd1 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:42.597 1+0 records in 00:14:42.597 1+0 records out 00:14:42.597 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000425748 s, 9.6 MB/s 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:42.597 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:42.598 05:42:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:42.857 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:42.857 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:42.857 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:42.857 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:42.857 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:42.857 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:42.857 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:42.857 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:42.857 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:42.857 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 94607 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 94607 ']' 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 94607 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 94607 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 94607' 00:14:43.116 killing process with pid 94607 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 94607 00:14:43.116 Received shutdown signal, test time was about 60.000000 seconds 00:14:43.116 00:14:43.116 Latency(us) 00:14:43.116 [2024-12-07T05:42:16.484Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:43.116 [2024-12-07T05:42:16.484Z] =================================================================================================================== 00:14:43.116 [2024-12-07T05:42:16.484Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:43.116 [2024-12-07 05:42:16.355819] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:43.116 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 94607 00:14:43.116 [2024-12-07 05:42:16.405552] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:14:43.376 00:14:43.376 real 0m18.084s 00:14:43.376 user 0m21.868s 00:14:43.376 sys 0m2.121s 00:14:43.376 ************************************ 00:14:43.376 END TEST raid5f_rebuild_test 00:14:43.376 ************************************ 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.376 05:42:16 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:14:43.376 05:42:16 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:43.376 05:42:16 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:43.376 05:42:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:43.376 ************************************ 00:14:43.376 START TEST raid5f_rebuild_test_sb 00:14:43.376 ************************************ 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 true false true 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:43.376 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=95107 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 95107 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 95107 ']' 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:43.377 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:43.377 05:42:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.636 [2024-12-07 05:42:16.791775] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:14:43.636 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:43.636 Zero copy mechanism will not be used. 00:14:43.636 [2024-12-07 05:42:16.792078] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95107 ] 00:14:43.636 [2024-12-07 05:42:16.953968] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:43.636 [2024-12-07 05:42:16.979954] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:43.896 [2024-12-07 05:42:17.022296] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:43.896 [2024-12-07 05:42:17.022406] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.466 BaseBdev1_malloc 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.466 [2024-12-07 05:42:17.645494] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:44.466 [2024-12-07 05:42:17.645601] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:44.466 [2024-12-07 05:42:17.645671] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:44.466 [2024-12-07 05:42:17.645684] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:44.466 [2024-12-07 05:42:17.647737] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:44.466 [2024-12-07 05:42:17.647774] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:44.466 BaseBdev1 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.466 BaseBdev2_malloc 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.466 [2024-12-07 05:42:17.673862] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:44.466 [2024-12-07 05:42:17.673916] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:44.466 [2024-12-07 05:42:17.673948] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:44.466 [2024-12-07 05:42:17.673956] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:44.466 [2024-12-07 05:42:17.675992] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:44.466 [2024-12-07 05:42:17.676084] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:44.466 BaseBdev2 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.466 BaseBdev3_malloc 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.466 [2024-12-07 05:42:17.702243] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:44.466 [2024-12-07 05:42:17.702297] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:44.466 [2024-12-07 05:42:17.702319] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:44.466 [2024-12-07 05:42:17.702328] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:44.466 [2024-12-07 05:42:17.704382] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:44.466 [2024-12-07 05:42:17.704420] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:44.466 BaseBdev3 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.466 BaseBdev4_malloc 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.466 [2024-12-07 05:42:17.741173] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:44.466 [2024-12-07 05:42:17.741222] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:44.466 [2024-12-07 05:42:17.741244] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:44.466 [2024-12-07 05:42:17.741252] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:44.466 [2024-12-07 05:42:17.743291] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:44.466 [2024-12-07 05:42:17.743330] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:44.466 BaseBdev4 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.466 spare_malloc 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.466 spare_delay 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.466 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.467 [2024-12-07 05:42:17.781612] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:44.467 [2024-12-07 05:42:17.781667] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:44.467 [2024-12-07 05:42:17.781702] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:44.467 [2024-12-07 05:42:17.781710] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:44.467 [2024-12-07 05:42:17.783829] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:44.467 [2024-12-07 05:42:17.783904] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:44.467 spare 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.467 [2024-12-07 05:42:17.793688] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:44.467 [2024-12-07 05:42:17.795558] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:44.467 [2024-12-07 05:42:17.795635] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:44.467 [2024-12-07 05:42:17.795705] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:44.467 [2024-12-07 05:42:17.795879] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:44.467 [2024-12-07 05:42:17.795896] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:44.467 [2024-12-07 05:42:17.796133] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:44.467 [2024-12-07 05:42:17.796554] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:44.467 [2024-12-07 05:42:17.796567] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:44.467 [2024-12-07 05:42:17.796696] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.467 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.727 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:44.727 "name": "raid_bdev1", 00:14:44.727 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:44.727 "strip_size_kb": 64, 00:14:44.727 "state": "online", 00:14:44.727 "raid_level": "raid5f", 00:14:44.727 "superblock": true, 00:14:44.727 "num_base_bdevs": 4, 00:14:44.727 "num_base_bdevs_discovered": 4, 00:14:44.727 "num_base_bdevs_operational": 4, 00:14:44.727 "base_bdevs_list": [ 00:14:44.727 { 00:14:44.727 "name": "BaseBdev1", 00:14:44.727 "uuid": "38155121-5008-533a-be7b-157f81c41c7a", 00:14:44.727 "is_configured": true, 00:14:44.727 "data_offset": 2048, 00:14:44.727 "data_size": 63488 00:14:44.727 }, 00:14:44.727 { 00:14:44.727 "name": "BaseBdev2", 00:14:44.727 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:44.727 "is_configured": true, 00:14:44.727 "data_offset": 2048, 00:14:44.727 "data_size": 63488 00:14:44.727 }, 00:14:44.727 { 00:14:44.727 "name": "BaseBdev3", 00:14:44.727 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:44.727 "is_configured": true, 00:14:44.727 "data_offset": 2048, 00:14:44.727 "data_size": 63488 00:14:44.727 }, 00:14:44.727 { 00:14:44.727 "name": "BaseBdev4", 00:14:44.727 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:44.727 "is_configured": true, 00:14:44.727 "data_offset": 2048, 00:14:44.727 "data_size": 63488 00:14:44.727 } 00:14:44.727 ] 00:14:44.727 }' 00:14:44.727 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:44.727 05:42:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:44.987 [2024-12-07 05:42:18.221878] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:44.987 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:44.988 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:44.988 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:44.988 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:44.988 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:44.988 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:44.988 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:44.988 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:44.988 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:44.988 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:44.988 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:44.988 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:45.247 [2024-12-07 05:42:18.501210] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:14:45.247 /dev/nbd0 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:45.247 1+0 records in 00:14:45.247 1+0 records out 00:14:45.247 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000204503 s, 20.0 MB/s 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:14:45.247 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:14:45.813 496+0 records in 00:14:45.813 496+0 records out 00:14:45.813 97517568 bytes (98 MB, 93 MiB) copied, 0.37408 s, 261 MB/s 00:14:45.813 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:45.813 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:45.813 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:45.813 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:45.813 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:45.813 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:45.813 05:42:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:45.813 [2024-12-07 05:42:19.137097] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:45.813 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:45.813 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:45.813 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:45.813 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:45.813 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:45.813 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:45.813 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:45.813 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:45.813 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:45.813 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.813 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.813 [2024-12-07 05:42:19.174740] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:46.077 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.077 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:46.077 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:46.077 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:46.077 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:46.078 "name": "raid_bdev1", 00:14:46.078 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:46.078 "strip_size_kb": 64, 00:14:46.078 "state": "online", 00:14:46.078 "raid_level": "raid5f", 00:14:46.078 "superblock": true, 00:14:46.078 "num_base_bdevs": 4, 00:14:46.078 "num_base_bdevs_discovered": 3, 00:14:46.078 "num_base_bdevs_operational": 3, 00:14:46.078 "base_bdevs_list": [ 00:14:46.078 { 00:14:46.078 "name": null, 00:14:46.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.078 "is_configured": false, 00:14:46.078 "data_offset": 0, 00:14:46.078 "data_size": 63488 00:14:46.078 }, 00:14:46.078 { 00:14:46.078 "name": "BaseBdev2", 00:14:46.078 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:46.078 "is_configured": true, 00:14:46.078 "data_offset": 2048, 00:14:46.078 "data_size": 63488 00:14:46.078 }, 00:14:46.078 { 00:14:46.078 "name": "BaseBdev3", 00:14:46.078 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:46.078 "is_configured": true, 00:14:46.078 "data_offset": 2048, 00:14:46.078 "data_size": 63488 00:14:46.078 }, 00:14:46.078 { 00:14:46.078 "name": "BaseBdev4", 00:14:46.078 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:46.078 "is_configured": true, 00:14:46.078 "data_offset": 2048, 00:14:46.078 "data_size": 63488 00:14:46.078 } 00:14:46.078 ] 00:14:46.078 }' 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:46.078 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.348 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:46.348 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.348 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.348 [2024-12-07 05:42:19.609965] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:46.348 [2024-12-07 05:42:19.614215] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000270a0 00:14:46.348 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.348 05:42:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:46.348 [2024-12-07 05:42:19.616493] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:47.293 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:47.293 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:47.293 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:47.293 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:47.293 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:47.293 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.293 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:47.293 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.293 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.293 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.551 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:47.551 "name": "raid_bdev1", 00:14:47.551 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:47.551 "strip_size_kb": 64, 00:14:47.551 "state": "online", 00:14:47.551 "raid_level": "raid5f", 00:14:47.551 "superblock": true, 00:14:47.551 "num_base_bdevs": 4, 00:14:47.551 "num_base_bdevs_discovered": 4, 00:14:47.551 "num_base_bdevs_operational": 4, 00:14:47.551 "process": { 00:14:47.551 "type": "rebuild", 00:14:47.551 "target": "spare", 00:14:47.551 "progress": { 00:14:47.551 "blocks": 19200, 00:14:47.551 "percent": 10 00:14:47.551 } 00:14:47.551 }, 00:14:47.551 "base_bdevs_list": [ 00:14:47.551 { 00:14:47.551 "name": "spare", 00:14:47.551 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:47.551 "is_configured": true, 00:14:47.551 "data_offset": 2048, 00:14:47.551 "data_size": 63488 00:14:47.551 }, 00:14:47.551 { 00:14:47.551 "name": "BaseBdev2", 00:14:47.551 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:47.551 "is_configured": true, 00:14:47.551 "data_offset": 2048, 00:14:47.551 "data_size": 63488 00:14:47.551 }, 00:14:47.551 { 00:14:47.551 "name": "BaseBdev3", 00:14:47.551 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:47.551 "is_configured": true, 00:14:47.551 "data_offset": 2048, 00:14:47.551 "data_size": 63488 00:14:47.551 }, 00:14:47.551 { 00:14:47.551 "name": "BaseBdev4", 00:14:47.551 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:47.551 "is_configured": true, 00:14:47.551 "data_offset": 2048, 00:14:47.551 "data_size": 63488 00:14:47.551 } 00:14:47.551 ] 00:14:47.551 }' 00:14:47.551 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:47.551 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:47.551 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:47.551 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:47.551 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:47.551 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.551 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.551 [2024-12-07 05:42:20.753038] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:47.551 [2024-12-07 05:42:20.822252] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:47.551 [2024-12-07 05:42:20.822327] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:47.551 [2024-12-07 05:42:20.822345] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:47.551 [2024-12-07 05:42:20.822355] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:47.551 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.552 "name": "raid_bdev1", 00:14:47.552 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:47.552 "strip_size_kb": 64, 00:14:47.552 "state": "online", 00:14:47.552 "raid_level": "raid5f", 00:14:47.552 "superblock": true, 00:14:47.552 "num_base_bdevs": 4, 00:14:47.552 "num_base_bdevs_discovered": 3, 00:14:47.552 "num_base_bdevs_operational": 3, 00:14:47.552 "base_bdevs_list": [ 00:14:47.552 { 00:14:47.552 "name": null, 00:14:47.552 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.552 "is_configured": false, 00:14:47.552 "data_offset": 0, 00:14:47.552 "data_size": 63488 00:14:47.552 }, 00:14:47.552 { 00:14:47.552 "name": "BaseBdev2", 00:14:47.552 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:47.552 "is_configured": true, 00:14:47.552 "data_offset": 2048, 00:14:47.552 "data_size": 63488 00:14:47.552 }, 00:14:47.552 { 00:14:47.552 "name": "BaseBdev3", 00:14:47.552 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:47.552 "is_configured": true, 00:14:47.552 "data_offset": 2048, 00:14:47.552 "data_size": 63488 00:14:47.552 }, 00:14:47.552 { 00:14:47.552 "name": "BaseBdev4", 00:14:47.552 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:47.552 "is_configured": true, 00:14:47.552 "data_offset": 2048, 00:14:47.552 "data_size": 63488 00:14:47.552 } 00:14:47.552 ] 00:14:47.552 }' 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.552 05:42:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:48.119 "name": "raid_bdev1", 00:14:48.119 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:48.119 "strip_size_kb": 64, 00:14:48.119 "state": "online", 00:14:48.119 "raid_level": "raid5f", 00:14:48.119 "superblock": true, 00:14:48.119 "num_base_bdevs": 4, 00:14:48.119 "num_base_bdevs_discovered": 3, 00:14:48.119 "num_base_bdevs_operational": 3, 00:14:48.119 "base_bdevs_list": [ 00:14:48.119 { 00:14:48.119 "name": null, 00:14:48.119 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.119 "is_configured": false, 00:14:48.119 "data_offset": 0, 00:14:48.119 "data_size": 63488 00:14:48.119 }, 00:14:48.119 { 00:14:48.119 "name": "BaseBdev2", 00:14:48.119 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:48.119 "is_configured": true, 00:14:48.119 "data_offset": 2048, 00:14:48.119 "data_size": 63488 00:14:48.119 }, 00:14:48.119 { 00:14:48.119 "name": "BaseBdev3", 00:14:48.119 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:48.119 "is_configured": true, 00:14:48.119 "data_offset": 2048, 00:14:48.119 "data_size": 63488 00:14:48.119 }, 00:14:48.119 { 00:14:48.119 "name": "BaseBdev4", 00:14:48.119 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:48.119 "is_configured": true, 00:14:48.119 "data_offset": 2048, 00:14:48.119 "data_size": 63488 00:14:48.119 } 00:14:48.119 ] 00:14:48.119 }' 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.119 [2024-12-07 05:42:21.379167] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:48.119 [2024-12-07 05:42:21.383367] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027170 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.119 05:42:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:48.119 [2024-12-07 05:42:21.385577] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:49.054 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:49.054 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:49.054 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:49.054 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:49.054 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:49.054 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.054 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:49.054 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.054 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.054 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.313 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:49.313 "name": "raid_bdev1", 00:14:49.313 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:49.313 "strip_size_kb": 64, 00:14:49.313 "state": "online", 00:14:49.313 "raid_level": "raid5f", 00:14:49.313 "superblock": true, 00:14:49.313 "num_base_bdevs": 4, 00:14:49.313 "num_base_bdevs_discovered": 4, 00:14:49.313 "num_base_bdevs_operational": 4, 00:14:49.313 "process": { 00:14:49.313 "type": "rebuild", 00:14:49.314 "target": "spare", 00:14:49.314 "progress": { 00:14:49.314 "blocks": 19200, 00:14:49.314 "percent": 10 00:14:49.314 } 00:14:49.314 }, 00:14:49.314 "base_bdevs_list": [ 00:14:49.314 { 00:14:49.314 "name": "spare", 00:14:49.314 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:49.314 "is_configured": true, 00:14:49.314 "data_offset": 2048, 00:14:49.314 "data_size": 63488 00:14:49.314 }, 00:14:49.314 { 00:14:49.314 "name": "BaseBdev2", 00:14:49.314 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:49.314 "is_configured": true, 00:14:49.314 "data_offset": 2048, 00:14:49.314 "data_size": 63488 00:14:49.314 }, 00:14:49.314 { 00:14:49.314 "name": "BaseBdev3", 00:14:49.314 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:49.314 "is_configured": true, 00:14:49.314 "data_offset": 2048, 00:14:49.314 "data_size": 63488 00:14:49.314 }, 00:14:49.314 { 00:14:49.314 "name": "BaseBdev4", 00:14:49.314 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:49.314 "is_configured": true, 00:14:49.314 "data_offset": 2048, 00:14:49.314 "data_size": 63488 00:14:49.314 } 00:14:49.314 ] 00:14:49.314 }' 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:14:49.314 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=519 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:49.314 "name": "raid_bdev1", 00:14:49.314 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:49.314 "strip_size_kb": 64, 00:14:49.314 "state": "online", 00:14:49.314 "raid_level": "raid5f", 00:14:49.314 "superblock": true, 00:14:49.314 "num_base_bdevs": 4, 00:14:49.314 "num_base_bdevs_discovered": 4, 00:14:49.314 "num_base_bdevs_operational": 4, 00:14:49.314 "process": { 00:14:49.314 "type": "rebuild", 00:14:49.314 "target": "spare", 00:14:49.314 "progress": { 00:14:49.314 "blocks": 21120, 00:14:49.314 "percent": 11 00:14:49.314 } 00:14:49.314 }, 00:14:49.314 "base_bdevs_list": [ 00:14:49.314 { 00:14:49.314 "name": "spare", 00:14:49.314 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:49.314 "is_configured": true, 00:14:49.314 "data_offset": 2048, 00:14:49.314 "data_size": 63488 00:14:49.314 }, 00:14:49.314 { 00:14:49.314 "name": "BaseBdev2", 00:14:49.314 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:49.314 "is_configured": true, 00:14:49.314 "data_offset": 2048, 00:14:49.314 "data_size": 63488 00:14:49.314 }, 00:14:49.314 { 00:14:49.314 "name": "BaseBdev3", 00:14:49.314 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:49.314 "is_configured": true, 00:14:49.314 "data_offset": 2048, 00:14:49.314 "data_size": 63488 00:14:49.314 }, 00:14:49.314 { 00:14:49.314 "name": "BaseBdev4", 00:14:49.314 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:49.314 "is_configured": true, 00:14:49.314 "data_offset": 2048, 00:14:49.314 "data_size": 63488 00:14:49.314 } 00:14:49.314 ] 00:14:49.314 }' 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:49.314 05:42:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:50.694 "name": "raid_bdev1", 00:14:50.694 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:50.694 "strip_size_kb": 64, 00:14:50.694 "state": "online", 00:14:50.694 "raid_level": "raid5f", 00:14:50.694 "superblock": true, 00:14:50.694 "num_base_bdevs": 4, 00:14:50.694 "num_base_bdevs_discovered": 4, 00:14:50.694 "num_base_bdevs_operational": 4, 00:14:50.694 "process": { 00:14:50.694 "type": "rebuild", 00:14:50.694 "target": "spare", 00:14:50.694 "progress": { 00:14:50.694 "blocks": 42240, 00:14:50.694 "percent": 22 00:14:50.694 } 00:14:50.694 }, 00:14:50.694 "base_bdevs_list": [ 00:14:50.694 { 00:14:50.694 "name": "spare", 00:14:50.694 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:50.694 "is_configured": true, 00:14:50.694 "data_offset": 2048, 00:14:50.694 "data_size": 63488 00:14:50.694 }, 00:14:50.694 { 00:14:50.694 "name": "BaseBdev2", 00:14:50.694 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:50.694 "is_configured": true, 00:14:50.694 "data_offset": 2048, 00:14:50.694 "data_size": 63488 00:14:50.694 }, 00:14:50.694 { 00:14:50.694 "name": "BaseBdev3", 00:14:50.694 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:50.694 "is_configured": true, 00:14:50.694 "data_offset": 2048, 00:14:50.694 "data_size": 63488 00:14:50.694 }, 00:14:50.694 { 00:14:50.694 "name": "BaseBdev4", 00:14:50.694 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:50.694 "is_configured": true, 00:14:50.694 "data_offset": 2048, 00:14:50.694 "data_size": 63488 00:14:50.694 } 00:14:50.694 ] 00:14:50.694 }' 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:50.694 05:42:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:51.635 "name": "raid_bdev1", 00:14:51.635 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:51.635 "strip_size_kb": 64, 00:14:51.635 "state": "online", 00:14:51.635 "raid_level": "raid5f", 00:14:51.635 "superblock": true, 00:14:51.635 "num_base_bdevs": 4, 00:14:51.635 "num_base_bdevs_discovered": 4, 00:14:51.635 "num_base_bdevs_operational": 4, 00:14:51.635 "process": { 00:14:51.635 "type": "rebuild", 00:14:51.635 "target": "spare", 00:14:51.635 "progress": { 00:14:51.635 "blocks": 63360, 00:14:51.635 "percent": 33 00:14:51.635 } 00:14:51.635 }, 00:14:51.635 "base_bdevs_list": [ 00:14:51.635 { 00:14:51.635 "name": "spare", 00:14:51.635 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:51.635 "is_configured": true, 00:14:51.635 "data_offset": 2048, 00:14:51.635 "data_size": 63488 00:14:51.635 }, 00:14:51.635 { 00:14:51.635 "name": "BaseBdev2", 00:14:51.635 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:51.635 "is_configured": true, 00:14:51.635 "data_offset": 2048, 00:14:51.635 "data_size": 63488 00:14:51.635 }, 00:14:51.635 { 00:14:51.635 "name": "BaseBdev3", 00:14:51.635 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:51.635 "is_configured": true, 00:14:51.635 "data_offset": 2048, 00:14:51.635 "data_size": 63488 00:14:51.635 }, 00:14:51.635 { 00:14:51.635 "name": "BaseBdev4", 00:14:51.635 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:51.635 "is_configured": true, 00:14:51.635 "data_offset": 2048, 00:14:51.635 "data_size": 63488 00:14:51.635 } 00:14:51.635 ] 00:14:51.635 }' 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:51.635 05:42:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:52.571 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:52.571 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:52.571 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:52.571 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:52.571 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:52.571 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:52.571 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.571 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.571 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:52.571 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.571 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.830 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:52.831 "name": "raid_bdev1", 00:14:52.831 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:52.831 "strip_size_kb": 64, 00:14:52.831 "state": "online", 00:14:52.831 "raid_level": "raid5f", 00:14:52.831 "superblock": true, 00:14:52.831 "num_base_bdevs": 4, 00:14:52.831 "num_base_bdevs_discovered": 4, 00:14:52.831 "num_base_bdevs_operational": 4, 00:14:52.831 "process": { 00:14:52.831 "type": "rebuild", 00:14:52.831 "target": "spare", 00:14:52.831 "progress": { 00:14:52.831 "blocks": 86400, 00:14:52.831 "percent": 45 00:14:52.831 } 00:14:52.831 }, 00:14:52.831 "base_bdevs_list": [ 00:14:52.831 { 00:14:52.831 "name": "spare", 00:14:52.831 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:52.831 "is_configured": true, 00:14:52.831 "data_offset": 2048, 00:14:52.831 "data_size": 63488 00:14:52.831 }, 00:14:52.831 { 00:14:52.831 "name": "BaseBdev2", 00:14:52.831 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:52.831 "is_configured": true, 00:14:52.831 "data_offset": 2048, 00:14:52.831 "data_size": 63488 00:14:52.831 }, 00:14:52.831 { 00:14:52.831 "name": "BaseBdev3", 00:14:52.831 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:52.831 "is_configured": true, 00:14:52.831 "data_offset": 2048, 00:14:52.831 "data_size": 63488 00:14:52.831 }, 00:14:52.831 { 00:14:52.831 "name": "BaseBdev4", 00:14:52.831 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:52.831 "is_configured": true, 00:14:52.831 "data_offset": 2048, 00:14:52.831 "data_size": 63488 00:14:52.831 } 00:14:52.831 ] 00:14:52.831 }' 00:14:52.831 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:52.831 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:52.831 05:42:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:52.831 05:42:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:52.831 05:42:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.771 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:53.771 "name": "raid_bdev1", 00:14:53.771 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:53.771 "strip_size_kb": 64, 00:14:53.771 "state": "online", 00:14:53.771 "raid_level": "raid5f", 00:14:53.771 "superblock": true, 00:14:53.771 "num_base_bdevs": 4, 00:14:53.771 "num_base_bdevs_discovered": 4, 00:14:53.771 "num_base_bdevs_operational": 4, 00:14:53.771 "process": { 00:14:53.771 "type": "rebuild", 00:14:53.771 "target": "spare", 00:14:53.771 "progress": { 00:14:53.771 "blocks": 107520, 00:14:53.772 "percent": 56 00:14:53.772 } 00:14:53.772 }, 00:14:53.772 "base_bdevs_list": [ 00:14:53.772 { 00:14:53.772 "name": "spare", 00:14:53.772 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:53.772 "is_configured": true, 00:14:53.772 "data_offset": 2048, 00:14:53.772 "data_size": 63488 00:14:53.772 }, 00:14:53.772 { 00:14:53.772 "name": "BaseBdev2", 00:14:53.772 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:53.772 "is_configured": true, 00:14:53.772 "data_offset": 2048, 00:14:53.772 "data_size": 63488 00:14:53.772 }, 00:14:53.772 { 00:14:53.772 "name": "BaseBdev3", 00:14:53.772 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:53.772 "is_configured": true, 00:14:53.772 "data_offset": 2048, 00:14:53.772 "data_size": 63488 00:14:53.772 }, 00:14:53.772 { 00:14:53.772 "name": "BaseBdev4", 00:14:53.772 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:53.772 "is_configured": true, 00:14:53.772 "data_offset": 2048, 00:14:53.772 "data_size": 63488 00:14:53.772 } 00:14:53.772 ] 00:14:53.772 }' 00:14:53.772 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:54.031 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:54.031 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:54.031 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:54.031 05:42:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:54.991 "name": "raid_bdev1", 00:14:54.991 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:54.991 "strip_size_kb": 64, 00:14:54.991 "state": "online", 00:14:54.991 "raid_level": "raid5f", 00:14:54.991 "superblock": true, 00:14:54.991 "num_base_bdevs": 4, 00:14:54.991 "num_base_bdevs_discovered": 4, 00:14:54.991 "num_base_bdevs_operational": 4, 00:14:54.991 "process": { 00:14:54.991 "type": "rebuild", 00:14:54.991 "target": "spare", 00:14:54.991 "progress": { 00:14:54.991 "blocks": 128640, 00:14:54.991 "percent": 67 00:14:54.991 } 00:14:54.991 }, 00:14:54.991 "base_bdevs_list": [ 00:14:54.991 { 00:14:54.991 "name": "spare", 00:14:54.991 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:54.991 "is_configured": true, 00:14:54.991 "data_offset": 2048, 00:14:54.991 "data_size": 63488 00:14:54.991 }, 00:14:54.991 { 00:14:54.991 "name": "BaseBdev2", 00:14:54.991 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:54.991 "is_configured": true, 00:14:54.991 "data_offset": 2048, 00:14:54.991 "data_size": 63488 00:14:54.991 }, 00:14:54.991 { 00:14:54.991 "name": "BaseBdev3", 00:14:54.991 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:54.991 "is_configured": true, 00:14:54.991 "data_offset": 2048, 00:14:54.991 "data_size": 63488 00:14:54.991 }, 00:14:54.991 { 00:14:54.991 "name": "BaseBdev4", 00:14:54.991 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:54.991 "is_configured": true, 00:14:54.991 "data_offset": 2048, 00:14:54.991 "data_size": 63488 00:14:54.991 } 00:14:54.991 ] 00:14:54.991 }' 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:54.991 05:42:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:56.376 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:56.376 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:56.376 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:56.376 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:56.376 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:56.376 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:56.376 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.376 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.376 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.376 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.377 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.377 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:56.377 "name": "raid_bdev1", 00:14:56.377 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:56.377 "strip_size_kb": 64, 00:14:56.377 "state": "online", 00:14:56.377 "raid_level": "raid5f", 00:14:56.377 "superblock": true, 00:14:56.377 "num_base_bdevs": 4, 00:14:56.377 "num_base_bdevs_discovered": 4, 00:14:56.377 "num_base_bdevs_operational": 4, 00:14:56.377 "process": { 00:14:56.377 "type": "rebuild", 00:14:56.377 "target": "spare", 00:14:56.377 "progress": { 00:14:56.377 "blocks": 151680, 00:14:56.377 "percent": 79 00:14:56.377 } 00:14:56.377 }, 00:14:56.377 "base_bdevs_list": [ 00:14:56.377 { 00:14:56.377 "name": "spare", 00:14:56.377 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:56.377 "is_configured": true, 00:14:56.377 "data_offset": 2048, 00:14:56.377 "data_size": 63488 00:14:56.377 }, 00:14:56.377 { 00:14:56.377 "name": "BaseBdev2", 00:14:56.377 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:56.377 "is_configured": true, 00:14:56.377 "data_offset": 2048, 00:14:56.377 "data_size": 63488 00:14:56.377 }, 00:14:56.377 { 00:14:56.377 "name": "BaseBdev3", 00:14:56.377 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:56.377 "is_configured": true, 00:14:56.377 "data_offset": 2048, 00:14:56.377 "data_size": 63488 00:14:56.377 }, 00:14:56.377 { 00:14:56.377 "name": "BaseBdev4", 00:14:56.377 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:56.377 "is_configured": true, 00:14:56.377 "data_offset": 2048, 00:14:56.377 "data_size": 63488 00:14:56.377 } 00:14:56.377 ] 00:14:56.377 }' 00:14:56.377 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:56.377 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:56.377 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:56.377 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:56.377 05:42:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:57.316 "name": "raid_bdev1", 00:14:57.316 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:57.316 "strip_size_kb": 64, 00:14:57.316 "state": "online", 00:14:57.316 "raid_level": "raid5f", 00:14:57.316 "superblock": true, 00:14:57.316 "num_base_bdevs": 4, 00:14:57.316 "num_base_bdevs_discovered": 4, 00:14:57.316 "num_base_bdevs_operational": 4, 00:14:57.316 "process": { 00:14:57.316 "type": "rebuild", 00:14:57.316 "target": "spare", 00:14:57.316 "progress": { 00:14:57.316 "blocks": 172800, 00:14:57.316 "percent": 90 00:14:57.316 } 00:14:57.316 }, 00:14:57.316 "base_bdevs_list": [ 00:14:57.316 { 00:14:57.316 "name": "spare", 00:14:57.316 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:57.316 "is_configured": true, 00:14:57.316 "data_offset": 2048, 00:14:57.316 "data_size": 63488 00:14:57.316 }, 00:14:57.316 { 00:14:57.316 "name": "BaseBdev2", 00:14:57.316 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:57.316 "is_configured": true, 00:14:57.316 "data_offset": 2048, 00:14:57.316 "data_size": 63488 00:14:57.316 }, 00:14:57.316 { 00:14:57.316 "name": "BaseBdev3", 00:14:57.316 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:57.316 "is_configured": true, 00:14:57.316 "data_offset": 2048, 00:14:57.316 "data_size": 63488 00:14:57.316 }, 00:14:57.316 { 00:14:57.316 "name": "BaseBdev4", 00:14:57.316 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:57.316 "is_configured": true, 00:14:57.316 "data_offset": 2048, 00:14:57.316 "data_size": 63488 00:14:57.316 } 00:14:57.316 ] 00:14:57.316 }' 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:57.316 05:42:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:58.254 [2024-12-07 05:42:31.427681] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:58.254 [2024-12-07 05:42:31.427851] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:58.254 [2024-12-07 05:42:31.428012] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.514 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:58.514 "name": "raid_bdev1", 00:14:58.514 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:58.514 "strip_size_kb": 64, 00:14:58.514 "state": "online", 00:14:58.514 "raid_level": "raid5f", 00:14:58.514 "superblock": true, 00:14:58.514 "num_base_bdevs": 4, 00:14:58.514 "num_base_bdevs_discovered": 4, 00:14:58.514 "num_base_bdevs_operational": 4, 00:14:58.514 "base_bdevs_list": [ 00:14:58.514 { 00:14:58.514 "name": "spare", 00:14:58.514 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:58.514 "is_configured": true, 00:14:58.514 "data_offset": 2048, 00:14:58.514 "data_size": 63488 00:14:58.514 }, 00:14:58.514 { 00:14:58.514 "name": "BaseBdev2", 00:14:58.514 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:58.514 "is_configured": true, 00:14:58.514 "data_offset": 2048, 00:14:58.514 "data_size": 63488 00:14:58.514 }, 00:14:58.514 { 00:14:58.515 "name": "BaseBdev3", 00:14:58.515 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:58.515 "is_configured": true, 00:14:58.515 "data_offset": 2048, 00:14:58.515 "data_size": 63488 00:14:58.515 }, 00:14:58.515 { 00:14:58.515 "name": "BaseBdev4", 00:14:58.515 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:58.515 "is_configured": true, 00:14:58.515 "data_offset": 2048, 00:14:58.515 "data_size": 63488 00:14:58.515 } 00:14:58.515 ] 00:14:58.515 }' 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:58.515 "name": "raid_bdev1", 00:14:58.515 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:58.515 "strip_size_kb": 64, 00:14:58.515 "state": "online", 00:14:58.515 "raid_level": "raid5f", 00:14:58.515 "superblock": true, 00:14:58.515 "num_base_bdevs": 4, 00:14:58.515 "num_base_bdevs_discovered": 4, 00:14:58.515 "num_base_bdevs_operational": 4, 00:14:58.515 "base_bdevs_list": [ 00:14:58.515 { 00:14:58.515 "name": "spare", 00:14:58.515 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:58.515 "is_configured": true, 00:14:58.515 "data_offset": 2048, 00:14:58.515 "data_size": 63488 00:14:58.515 }, 00:14:58.515 { 00:14:58.515 "name": "BaseBdev2", 00:14:58.515 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:58.515 "is_configured": true, 00:14:58.515 "data_offset": 2048, 00:14:58.515 "data_size": 63488 00:14:58.515 }, 00:14:58.515 { 00:14:58.515 "name": "BaseBdev3", 00:14:58.515 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:58.515 "is_configured": true, 00:14:58.515 "data_offset": 2048, 00:14:58.515 "data_size": 63488 00:14:58.515 }, 00:14:58.515 { 00:14:58.515 "name": "BaseBdev4", 00:14:58.515 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:58.515 "is_configured": true, 00:14:58.515 "data_offset": 2048, 00:14:58.515 "data_size": 63488 00:14:58.515 } 00:14:58.515 ] 00:14:58.515 }' 00:14:58.515 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:58.774 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:58.774 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:58.774 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:58.774 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:58.775 "name": "raid_bdev1", 00:14:58.775 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:14:58.775 "strip_size_kb": 64, 00:14:58.775 "state": "online", 00:14:58.775 "raid_level": "raid5f", 00:14:58.775 "superblock": true, 00:14:58.775 "num_base_bdevs": 4, 00:14:58.775 "num_base_bdevs_discovered": 4, 00:14:58.775 "num_base_bdevs_operational": 4, 00:14:58.775 "base_bdevs_list": [ 00:14:58.775 { 00:14:58.775 "name": "spare", 00:14:58.775 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:14:58.775 "is_configured": true, 00:14:58.775 "data_offset": 2048, 00:14:58.775 "data_size": 63488 00:14:58.775 }, 00:14:58.775 { 00:14:58.775 "name": "BaseBdev2", 00:14:58.775 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:14:58.775 "is_configured": true, 00:14:58.775 "data_offset": 2048, 00:14:58.775 "data_size": 63488 00:14:58.775 }, 00:14:58.775 { 00:14:58.775 "name": "BaseBdev3", 00:14:58.775 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:14:58.775 "is_configured": true, 00:14:58.775 "data_offset": 2048, 00:14:58.775 "data_size": 63488 00:14:58.775 }, 00:14:58.775 { 00:14:58.775 "name": "BaseBdev4", 00:14:58.775 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:14:58.775 "is_configured": true, 00:14:58.775 "data_offset": 2048, 00:14:58.775 "data_size": 63488 00:14:58.775 } 00:14:58.775 ] 00:14:58.775 }' 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:58.775 05:42:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.034 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:59.034 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.034 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.034 [2024-12-07 05:42:32.395939] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:59.034 [2024-12-07 05:42:32.396027] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:59.034 [2024-12-07 05:42:32.396131] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:59.034 [2024-12-07 05:42:32.396266] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:59.034 [2024-12-07 05:42:32.396323] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:59.034 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:59.295 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:59.295 /dev/nbd0 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:59.555 1+0 records in 00:14:59.555 1+0 records out 00:14:59.555 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000486755 s, 8.4 MB/s 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:59.555 /dev/nbd1 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:59.555 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:59.815 1+0 records in 00:14:59.815 1+0 records out 00:14:59.815 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000472087 s, 8.7 MB/s 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:59.815 05:42:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:59.815 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:59.815 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:59.815 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:59.815 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:59.815 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:59.815 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:59.815 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:00.075 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.335 [2024-12-07 05:42:33.464331] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:00.335 [2024-12-07 05:42:33.464403] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:00.335 [2024-12-07 05:42:33.464427] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:15:00.335 [2024-12-07 05:42:33.464438] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:00.335 [2024-12-07 05:42:33.466715] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:00.335 [2024-12-07 05:42:33.466811] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:00.335 [2024-12-07 05:42:33.466916] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:00.335 [2024-12-07 05:42:33.466986] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:00.335 [2024-12-07 05:42:33.467140] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:00.335 [2024-12-07 05:42:33.467273] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:00.335 [2024-12-07 05:42:33.467387] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:00.335 spare 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.335 [2024-12-07 05:42:33.567332] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:00.335 [2024-12-07 05:42:33.567358] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:00.335 [2024-12-07 05:42:33.567645] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000045820 00:15:00.335 [2024-12-07 05:42:33.568130] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:00.335 [2024-12-07 05:42:33.568153] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:00.335 [2024-12-07 05:42:33.568309] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:00.335 "name": "raid_bdev1", 00:15:00.335 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:00.335 "strip_size_kb": 64, 00:15:00.335 "state": "online", 00:15:00.335 "raid_level": "raid5f", 00:15:00.335 "superblock": true, 00:15:00.335 "num_base_bdevs": 4, 00:15:00.335 "num_base_bdevs_discovered": 4, 00:15:00.335 "num_base_bdevs_operational": 4, 00:15:00.335 "base_bdevs_list": [ 00:15:00.335 { 00:15:00.335 "name": "spare", 00:15:00.335 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:15:00.335 "is_configured": true, 00:15:00.335 "data_offset": 2048, 00:15:00.335 "data_size": 63488 00:15:00.335 }, 00:15:00.335 { 00:15:00.335 "name": "BaseBdev2", 00:15:00.335 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:00.335 "is_configured": true, 00:15:00.335 "data_offset": 2048, 00:15:00.335 "data_size": 63488 00:15:00.335 }, 00:15:00.335 { 00:15:00.335 "name": "BaseBdev3", 00:15:00.335 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:00.335 "is_configured": true, 00:15:00.335 "data_offset": 2048, 00:15:00.335 "data_size": 63488 00:15:00.335 }, 00:15:00.335 { 00:15:00.335 "name": "BaseBdev4", 00:15:00.335 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:00.335 "is_configured": true, 00:15:00.335 "data_offset": 2048, 00:15:00.335 "data_size": 63488 00:15:00.335 } 00:15:00.335 ] 00:15:00.335 }' 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:00.335 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.905 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:00.905 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:00.905 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:00.905 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:00.905 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:00.905 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.905 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.905 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.905 05:42:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:00.905 "name": "raid_bdev1", 00:15:00.905 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:00.905 "strip_size_kb": 64, 00:15:00.905 "state": "online", 00:15:00.905 "raid_level": "raid5f", 00:15:00.905 "superblock": true, 00:15:00.905 "num_base_bdevs": 4, 00:15:00.905 "num_base_bdevs_discovered": 4, 00:15:00.905 "num_base_bdevs_operational": 4, 00:15:00.905 "base_bdevs_list": [ 00:15:00.905 { 00:15:00.905 "name": "spare", 00:15:00.905 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:15:00.905 "is_configured": true, 00:15:00.905 "data_offset": 2048, 00:15:00.905 "data_size": 63488 00:15:00.905 }, 00:15:00.905 { 00:15:00.905 "name": "BaseBdev2", 00:15:00.905 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:00.905 "is_configured": true, 00:15:00.905 "data_offset": 2048, 00:15:00.905 "data_size": 63488 00:15:00.905 }, 00:15:00.905 { 00:15:00.905 "name": "BaseBdev3", 00:15:00.905 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:00.905 "is_configured": true, 00:15:00.905 "data_offset": 2048, 00:15:00.905 "data_size": 63488 00:15:00.905 }, 00:15:00.905 { 00:15:00.905 "name": "BaseBdev4", 00:15:00.905 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:00.905 "is_configured": true, 00:15:00.905 "data_offset": 2048, 00:15:00.905 "data_size": 63488 00:15:00.905 } 00:15:00.905 ] 00:15:00.905 }' 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.905 [2024-12-07 05:42:34.159274] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:00.905 "name": "raid_bdev1", 00:15:00.905 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:00.905 "strip_size_kb": 64, 00:15:00.905 "state": "online", 00:15:00.905 "raid_level": "raid5f", 00:15:00.905 "superblock": true, 00:15:00.905 "num_base_bdevs": 4, 00:15:00.905 "num_base_bdevs_discovered": 3, 00:15:00.905 "num_base_bdevs_operational": 3, 00:15:00.905 "base_bdevs_list": [ 00:15:00.905 { 00:15:00.905 "name": null, 00:15:00.905 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:00.905 "is_configured": false, 00:15:00.905 "data_offset": 0, 00:15:00.905 "data_size": 63488 00:15:00.905 }, 00:15:00.905 { 00:15:00.905 "name": "BaseBdev2", 00:15:00.905 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:00.905 "is_configured": true, 00:15:00.905 "data_offset": 2048, 00:15:00.905 "data_size": 63488 00:15:00.905 }, 00:15:00.905 { 00:15:00.905 "name": "BaseBdev3", 00:15:00.905 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:00.905 "is_configured": true, 00:15:00.905 "data_offset": 2048, 00:15:00.905 "data_size": 63488 00:15:00.905 }, 00:15:00.905 { 00:15:00.905 "name": "BaseBdev4", 00:15:00.905 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:00.905 "is_configured": true, 00:15:00.905 "data_offset": 2048, 00:15:00.905 "data_size": 63488 00:15:00.905 } 00:15:00.905 ] 00:15:00.905 }' 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:00.905 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.475 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:01.475 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.475 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.475 [2024-12-07 05:42:34.562620] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:01.475 [2024-12-07 05:42:34.562798] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:01.475 [2024-12-07 05:42:34.562821] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:01.475 [2024-12-07 05:42:34.562872] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:01.475 [2024-12-07 05:42:34.566797] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000458f0 00:15:01.475 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.475 05:42:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:01.475 [2024-12-07 05:42:34.568955] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:02.412 "name": "raid_bdev1", 00:15:02.412 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:02.412 "strip_size_kb": 64, 00:15:02.412 "state": "online", 00:15:02.412 "raid_level": "raid5f", 00:15:02.412 "superblock": true, 00:15:02.412 "num_base_bdevs": 4, 00:15:02.412 "num_base_bdevs_discovered": 4, 00:15:02.412 "num_base_bdevs_operational": 4, 00:15:02.412 "process": { 00:15:02.412 "type": "rebuild", 00:15:02.412 "target": "spare", 00:15:02.412 "progress": { 00:15:02.412 "blocks": 19200, 00:15:02.412 "percent": 10 00:15:02.412 } 00:15:02.412 }, 00:15:02.412 "base_bdevs_list": [ 00:15:02.412 { 00:15:02.412 "name": "spare", 00:15:02.412 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:15:02.412 "is_configured": true, 00:15:02.412 "data_offset": 2048, 00:15:02.412 "data_size": 63488 00:15:02.412 }, 00:15:02.412 { 00:15:02.412 "name": "BaseBdev2", 00:15:02.412 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:02.412 "is_configured": true, 00:15:02.412 "data_offset": 2048, 00:15:02.412 "data_size": 63488 00:15:02.412 }, 00:15:02.412 { 00:15:02.412 "name": "BaseBdev3", 00:15:02.412 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:02.412 "is_configured": true, 00:15:02.412 "data_offset": 2048, 00:15:02.412 "data_size": 63488 00:15:02.412 }, 00:15:02.412 { 00:15:02.412 "name": "BaseBdev4", 00:15:02.412 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:02.412 "is_configured": true, 00:15:02.412 "data_offset": 2048, 00:15:02.412 "data_size": 63488 00:15:02.412 } 00:15:02.412 ] 00:15:02.412 }' 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:02.412 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:02.413 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:02.413 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:02.413 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.413 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.413 [2024-12-07 05:42:35.725407] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:02.413 [2024-12-07 05:42:35.774394] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:02.413 [2024-12-07 05:42:35.774455] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:02.413 [2024-12-07 05:42:35.774475] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:02.413 [2024-12-07 05:42:35.774483] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:02.672 "name": "raid_bdev1", 00:15:02.672 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:02.672 "strip_size_kb": 64, 00:15:02.672 "state": "online", 00:15:02.672 "raid_level": "raid5f", 00:15:02.672 "superblock": true, 00:15:02.672 "num_base_bdevs": 4, 00:15:02.672 "num_base_bdevs_discovered": 3, 00:15:02.672 "num_base_bdevs_operational": 3, 00:15:02.672 "base_bdevs_list": [ 00:15:02.672 { 00:15:02.672 "name": null, 00:15:02.672 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:02.672 "is_configured": false, 00:15:02.672 "data_offset": 0, 00:15:02.672 "data_size": 63488 00:15:02.672 }, 00:15:02.672 { 00:15:02.672 "name": "BaseBdev2", 00:15:02.672 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:02.672 "is_configured": true, 00:15:02.672 "data_offset": 2048, 00:15:02.672 "data_size": 63488 00:15:02.672 }, 00:15:02.672 { 00:15:02.672 "name": "BaseBdev3", 00:15:02.672 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:02.672 "is_configured": true, 00:15:02.672 "data_offset": 2048, 00:15:02.672 "data_size": 63488 00:15:02.672 }, 00:15:02.672 { 00:15:02.672 "name": "BaseBdev4", 00:15:02.672 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:02.672 "is_configured": true, 00:15:02.672 "data_offset": 2048, 00:15:02.672 "data_size": 63488 00:15:02.672 } 00:15:02.672 ] 00:15:02.672 }' 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:02.672 05:42:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.931 05:42:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:02.931 05:42:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.931 05:42:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.931 [2024-12-07 05:42:36.243005] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:02.931 [2024-12-07 05:42:36.243068] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:02.931 [2024-12-07 05:42:36.243102] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:15:02.931 [2024-12-07 05:42:36.243114] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:02.931 [2024-12-07 05:42:36.243544] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:02.931 [2024-12-07 05:42:36.243571] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:02.931 [2024-12-07 05:42:36.243671] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:02.931 [2024-12-07 05:42:36.243686] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:02.931 [2024-12-07 05:42:36.243702] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:02.931 [2024-12-07 05:42:36.243722] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:02.931 [2024-12-07 05:42:36.247561] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000459c0 00:15:02.931 spare 00:15:02.931 05:42:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.931 05:42:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:02.931 [2024-12-07 05:42:36.249736] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:04.313 "name": "raid_bdev1", 00:15:04.313 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:04.313 "strip_size_kb": 64, 00:15:04.313 "state": "online", 00:15:04.313 "raid_level": "raid5f", 00:15:04.313 "superblock": true, 00:15:04.313 "num_base_bdevs": 4, 00:15:04.313 "num_base_bdevs_discovered": 4, 00:15:04.313 "num_base_bdevs_operational": 4, 00:15:04.313 "process": { 00:15:04.313 "type": "rebuild", 00:15:04.313 "target": "spare", 00:15:04.313 "progress": { 00:15:04.313 "blocks": 19200, 00:15:04.313 "percent": 10 00:15:04.313 } 00:15:04.313 }, 00:15:04.313 "base_bdevs_list": [ 00:15:04.313 { 00:15:04.313 "name": "spare", 00:15:04.313 "uuid": "824d5b38-1585-5b2c-874f-2b1ff80b2351", 00:15:04.313 "is_configured": true, 00:15:04.313 "data_offset": 2048, 00:15:04.313 "data_size": 63488 00:15:04.313 }, 00:15:04.313 { 00:15:04.313 "name": "BaseBdev2", 00:15:04.313 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:04.313 "is_configured": true, 00:15:04.313 "data_offset": 2048, 00:15:04.313 "data_size": 63488 00:15:04.313 }, 00:15:04.313 { 00:15:04.313 "name": "BaseBdev3", 00:15:04.313 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:04.313 "is_configured": true, 00:15:04.313 "data_offset": 2048, 00:15:04.313 "data_size": 63488 00:15:04.313 }, 00:15:04.313 { 00:15:04.313 "name": "BaseBdev4", 00:15:04.313 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:04.313 "is_configured": true, 00:15:04.313 "data_offset": 2048, 00:15:04.313 "data_size": 63488 00:15:04.313 } 00:15:04.313 ] 00:15:04.313 }' 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.313 [2024-12-07 05:42:37.390527] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:04.313 [2024-12-07 05:42:37.455059] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:04.313 [2024-12-07 05:42:37.455121] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:04.313 [2024-12-07 05:42:37.455152] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:04.313 [2024-12-07 05:42:37.455161] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.313 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:04.313 "name": "raid_bdev1", 00:15:04.313 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:04.313 "strip_size_kb": 64, 00:15:04.313 "state": "online", 00:15:04.313 "raid_level": "raid5f", 00:15:04.313 "superblock": true, 00:15:04.313 "num_base_bdevs": 4, 00:15:04.313 "num_base_bdevs_discovered": 3, 00:15:04.313 "num_base_bdevs_operational": 3, 00:15:04.313 "base_bdevs_list": [ 00:15:04.313 { 00:15:04.313 "name": null, 00:15:04.313 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:04.313 "is_configured": false, 00:15:04.313 "data_offset": 0, 00:15:04.313 "data_size": 63488 00:15:04.313 }, 00:15:04.313 { 00:15:04.313 "name": "BaseBdev2", 00:15:04.313 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:04.313 "is_configured": true, 00:15:04.313 "data_offset": 2048, 00:15:04.313 "data_size": 63488 00:15:04.313 }, 00:15:04.313 { 00:15:04.313 "name": "BaseBdev3", 00:15:04.314 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:04.314 "is_configured": true, 00:15:04.314 "data_offset": 2048, 00:15:04.314 "data_size": 63488 00:15:04.314 }, 00:15:04.314 { 00:15:04.314 "name": "BaseBdev4", 00:15:04.314 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:04.314 "is_configured": true, 00:15:04.314 "data_offset": 2048, 00:15:04.314 "data_size": 63488 00:15:04.314 } 00:15:04.314 ] 00:15:04.314 }' 00:15:04.314 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:04.314 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.573 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:04.573 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:04.573 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:04.573 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:04.573 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:04.573 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.573 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:04.573 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.573 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.573 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.833 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:04.833 "name": "raid_bdev1", 00:15:04.833 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:04.833 "strip_size_kb": 64, 00:15:04.833 "state": "online", 00:15:04.833 "raid_level": "raid5f", 00:15:04.833 "superblock": true, 00:15:04.833 "num_base_bdevs": 4, 00:15:04.833 "num_base_bdevs_discovered": 3, 00:15:04.833 "num_base_bdevs_operational": 3, 00:15:04.833 "base_bdevs_list": [ 00:15:04.833 { 00:15:04.833 "name": null, 00:15:04.833 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:04.833 "is_configured": false, 00:15:04.833 "data_offset": 0, 00:15:04.833 "data_size": 63488 00:15:04.833 }, 00:15:04.833 { 00:15:04.833 "name": "BaseBdev2", 00:15:04.833 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:04.833 "is_configured": true, 00:15:04.833 "data_offset": 2048, 00:15:04.833 "data_size": 63488 00:15:04.833 }, 00:15:04.833 { 00:15:04.833 "name": "BaseBdev3", 00:15:04.833 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:04.833 "is_configured": true, 00:15:04.833 "data_offset": 2048, 00:15:04.833 "data_size": 63488 00:15:04.833 }, 00:15:04.833 { 00:15:04.833 "name": "BaseBdev4", 00:15:04.833 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:04.833 "is_configured": true, 00:15:04.833 "data_offset": 2048, 00:15:04.833 "data_size": 63488 00:15:04.833 } 00:15:04.833 ] 00:15:04.833 }' 00:15:04.833 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:04.833 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:04.833 05:42:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:04.833 05:42:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:04.833 05:42:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:04.833 05:42:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.833 05:42:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.833 05:42:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.833 05:42:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:04.833 05:42:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.833 05:42:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.833 [2024-12-07 05:42:38.059378] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:04.833 [2024-12-07 05:42:38.059440] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:04.833 [2024-12-07 05:42:38.059460] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:15:04.834 [2024-12-07 05:42:38.059470] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:04.834 [2024-12-07 05:42:38.059881] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:04.834 [2024-12-07 05:42:38.059914] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:04.834 [2024-12-07 05:42:38.059985] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:04.834 [2024-12-07 05:42:38.060008] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:04.834 [2024-12-07 05:42:38.060016] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:04.834 [2024-12-07 05:42:38.060029] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:04.834 BaseBdev1 00:15:04.834 05:42:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.834 05:42:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:05.772 "name": "raid_bdev1", 00:15:05.772 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:05.772 "strip_size_kb": 64, 00:15:05.772 "state": "online", 00:15:05.772 "raid_level": "raid5f", 00:15:05.772 "superblock": true, 00:15:05.772 "num_base_bdevs": 4, 00:15:05.772 "num_base_bdevs_discovered": 3, 00:15:05.772 "num_base_bdevs_operational": 3, 00:15:05.772 "base_bdevs_list": [ 00:15:05.772 { 00:15:05.772 "name": null, 00:15:05.772 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:05.772 "is_configured": false, 00:15:05.772 "data_offset": 0, 00:15:05.772 "data_size": 63488 00:15:05.772 }, 00:15:05.772 { 00:15:05.772 "name": "BaseBdev2", 00:15:05.772 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:05.772 "is_configured": true, 00:15:05.772 "data_offset": 2048, 00:15:05.772 "data_size": 63488 00:15:05.772 }, 00:15:05.772 { 00:15:05.772 "name": "BaseBdev3", 00:15:05.772 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:05.772 "is_configured": true, 00:15:05.772 "data_offset": 2048, 00:15:05.772 "data_size": 63488 00:15:05.772 }, 00:15:05.772 { 00:15:05.772 "name": "BaseBdev4", 00:15:05.772 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:05.772 "is_configured": true, 00:15:05.772 "data_offset": 2048, 00:15:05.772 "data_size": 63488 00:15:05.772 } 00:15:05.772 ] 00:15:05.772 }' 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:05.772 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.338 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:06.338 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:06.338 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:06.338 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:06.339 "name": "raid_bdev1", 00:15:06.339 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:06.339 "strip_size_kb": 64, 00:15:06.339 "state": "online", 00:15:06.339 "raid_level": "raid5f", 00:15:06.339 "superblock": true, 00:15:06.339 "num_base_bdevs": 4, 00:15:06.339 "num_base_bdevs_discovered": 3, 00:15:06.339 "num_base_bdevs_operational": 3, 00:15:06.339 "base_bdevs_list": [ 00:15:06.339 { 00:15:06.339 "name": null, 00:15:06.339 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:06.339 "is_configured": false, 00:15:06.339 "data_offset": 0, 00:15:06.339 "data_size": 63488 00:15:06.339 }, 00:15:06.339 { 00:15:06.339 "name": "BaseBdev2", 00:15:06.339 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:06.339 "is_configured": true, 00:15:06.339 "data_offset": 2048, 00:15:06.339 "data_size": 63488 00:15:06.339 }, 00:15:06.339 { 00:15:06.339 "name": "BaseBdev3", 00:15:06.339 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:06.339 "is_configured": true, 00:15:06.339 "data_offset": 2048, 00:15:06.339 "data_size": 63488 00:15:06.339 }, 00:15:06.339 { 00:15:06.339 "name": "BaseBdev4", 00:15:06.339 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:06.339 "is_configured": true, 00:15:06.339 "data_offset": 2048, 00:15:06.339 "data_size": 63488 00:15:06.339 } 00:15:06.339 ] 00:15:06.339 }' 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.339 [2024-12-07 05:42:39.664742] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:06.339 [2024-12-07 05:42:39.664926] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:06.339 [2024-12-07 05:42:39.664954] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:06.339 request: 00:15:06.339 { 00:15:06.339 "base_bdev": "BaseBdev1", 00:15:06.339 "raid_bdev": "raid_bdev1", 00:15:06.339 "method": "bdev_raid_add_base_bdev", 00:15:06.339 "req_id": 1 00:15:06.339 } 00:15:06.339 Got JSON-RPC error response 00:15:06.339 response: 00:15:06.339 { 00:15:06.339 "code": -22, 00:15:06.339 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:06.339 } 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:06.339 05:42:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:07.718 "name": "raid_bdev1", 00:15:07.718 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:07.718 "strip_size_kb": 64, 00:15:07.718 "state": "online", 00:15:07.718 "raid_level": "raid5f", 00:15:07.718 "superblock": true, 00:15:07.718 "num_base_bdevs": 4, 00:15:07.718 "num_base_bdevs_discovered": 3, 00:15:07.718 "num_base_bdevs_operational": 3, 00:15:07.718 "base_bdevs_list": [ 00:15:07.718 { 00:15:07.718 "name": null, 00:15:07.718 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:07.718 "is_configured": false, 00:15:07.718 "data_offset": 0, 00:15:07.718 "data_size": 63488 00:15:07.718 }, 00:15:07.718 { 00:15:07.718 "name": "BaseBdev2", 00:15:07.718 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:07.718 "is_configured": true, 00:15:07.718 "data_offset": 2048, 00:15:07.718 "data_size": 63488 00:15:07.718 }, 00:15:07.718 { 00:15:07.718 "name": "BaseBdev3", 00:15:07.718 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:07.718 "is_configured": true, 00:15:07.718 "data_offset": 2048, 00:15:07.718 "data_size": 63488 00:15:07.718 }, 00:15:07.718 { 00:15:07.718 "name": "BaseBdev4", 00:15:07.718 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:07.718 "is_configured": true, 00:15:07.718 "data_offset": 2048, 00:15:07.718 "data_size": 63488 00:15:07.718 } 00:15:07.718 ] 00:15:07.718 }' 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:07.718 05:42:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:07.978 "name": "raid_bdev1", 00:15:07.978 "uuid": "e7df6d77-03be-4df2-be8e-fa6271d6315f", 00:15:07.978 "strip_size_kb": 64, 00:15:07.978 "state": "online", 00:15:07.978 "raid_level": "raid5f", 00:15:07.978 "superblock": true, 00:15:07.978 "num_base_bdevs": 4, 00:15:07.978 "num_base_bdevs_discovered": 3, 00:15:07.978 "num_base_bdevs_operational": 3, 00:15:07.978 "base_bdevs_list": [ 00:15:07.978 { 00:15:07.978 "name": null, 00:15:07.978 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:07.978 "is_configured": false, 00:15:07.978 "data_offset": 0, 00:15:07.978 "data_size": 63488 00:15:07.978 }, 00:15:07.978 { 00:15:07.978 "name": "BaseBdev2", 00:15:07.978 "uuid": "8df439a9-652e-59b9-828e-abdc164b6d98", 00:15:07.978 "is_configured": true, 00:15:07.978 "data_offset": 2048, 00:15:07.978 "data_size": 63488 00:15:07.978 }, 00:15:07.978 { 00:15:07.978 "name": "BaseBdev3", 00:15:07.978 "uuid": "bb2d7b2b-cea7-5644-97e5-af5de5ac291e", 00:15:07.978 "is_configured": true, 00:15:07.978 "data_offset": 2048, 00:15:07.978 "data_size": 63488 00:15:07.978 }, 00:15:07.978 { 00:15:07.978 "name": "BaseBdev4", 00:15:07.978 "uuid": "ab5301c1-e71d-57fd-8128-157cf25bf5a9", 00:15:07.978 "is_configured": true, 00:15:07.978 "data_offset": 2048, 00:15:07.978 "data_size": 63488 00:15:07.978 } 00:15:07.978 ] 00:15:07.978 }' 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 95107 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 95107 ']' 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 95107 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 95107 00:15:07.978 killing process with pid 95107 00:15:07.978 Received shutdown signal, test time was about 60.000000 seconds 00:15:07.978 00:15:07.978 Latency(us) 00:15:07.978 [2024-12-07T05:42:41.346Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:07.978 [2024-12-07T05:42:41.346Z] =================================================================================================================== 00:15:07.978 [2024-12-07T05:42:41.346Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 95107' 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 95107 00:15:07.978 [2024-12-07 05:42:41.246166] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:07.978 [2024-12-07 05:42:41.246290] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:07.978 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 95107 00:15:07.978 [2024-12-07 05:42:41.246367] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:07.978 [2024-12-07 05:42:41.246388] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:15:07.978 [2024-12-07 05:42:41.297008] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:08.238 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:15:08.238 00:15:08.238 real 0m24.818s 00:15:08.238 user 0m31.457s 00:15:08.238 sys 0m2.885s 00:15:08.238 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:08.238 ************************************ 00:15:08.238 END TEST raid5f_rebuild_test_sb 00:15:08.238 ************************************ 00:15:08.238 05:42:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.238 05:42:41 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:15:08.238 05:42:41 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:15:08.238 05:42:41 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:15:08.238 05:42:41 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:08.238 05:42:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:08.238 ************************************ 00:15:08.238 START TEST raid_state_function_test_sb_4k 00:15:08.238 ************************************ 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=95899 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:08.238 Process raid pid: 95899 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 95899' 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 95899 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 95899 ']' 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:08.238 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:08.238 05:42:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:08.498 [2024-12-07 05:42:41.656775] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:15:08.498 [2024-12-07 05:42:41.656897] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:08.498 [2024-12-07 05:42:41.811675] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:08.498 [2024-12-07 05:42:41.835960] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:08.757 [2024-12-07 05:42:41.877507] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:08.757 [2024-12-07 05:42:41.877552] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:09.325 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:09.325 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:15:09.325 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:09.325 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.325 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:09.325 [2024-12-07 05:42:42.483713] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:09.325 [2024-12-07 05:42:42.483778] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:09.325 [2024-12-07 05:42:42.483788] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:09.326 [2024-12-07 05:42:42.483815] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:09.326 "name": "Existed_Raid", 00:15:09.326 "uuid": "d46ba527-a9e3-40ab-ac85-6be11453975a", 00:15:09.326 "strip_size_kb": 0, 00:15:09.326 "state": "configuring", 00:15:09.326 "raid_level": "raid1", 00:15:09.326 "superblock": true, 00:15:09.326 "num_base_bdevs": 2, 00:15:09.326 "num_base_bdevs_discovered": 0, 00:15:09.326 "num_base_bdevs_operational": 2, 00:15:09.326 "base_bdevs_list": [ 00:15:09.326 { 00:15:09.326 "name": "BaseBdev1", 00:15:09.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:09.326 "is_configured": false, 00:15:09.326 "data_offset": 0, 00:15:09.326 "data_size": 0 00:15:09.326 }, 00:15:09.326 { 00:15:09.326 "name": "BaseBdev2", 00:15:09.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:09.326 "is_configured": false, 00:15:09.326 "data_offset": 0, 00:15:09.326 "data_size": 0 00:15:09.326 } 00:15:09.326 ] 00:15:09.326 }' 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:09.326 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:09.586 [2024-12-07 05:42:42.898874] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:09.586 [2024-12-07 05:42:42.898916] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:09.586 [2024-12-07 05:42:42.910865] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:09.586 [2024-12-07 05:42:42.910924] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:09.586 [2024-12-07 05:42:42.910932] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:09.586 [2024-12-07 05:42:42.910950] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:09.586 [2024-12-07 05:42:42.931519] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:09.586 BaseBdev1 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.586 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:09.847 [ 00:15:09.847 { 00:15:09.847 "name": "BaseBdev1", 00:15:09.847 "aliases": [ 00:15:09.847 "faf6d498-d622-46fd-9cdd-aee5b2b35c23" 00:15:09.847 ], 00:15:09.847 "product_name": "Malloc disk", 00:15:09.847 "block_size": 4096, 00:15:09.847 "num_blocks": 8192, 00:15:09.847 "uuid": "faf6d498-d622-46fd-9cdd-aee5b2b35c23", 00:15:09.847 "assigned_rate_limits": { 00:15:09.847 "rw_ios_per_sec": 0, 00:15:09.847 "rw_mbytes_per_sec": 0, 00:15:09.847 "r_mbytes_per_sec": 0, 00:15:09.847 "w_mbytes_per_sec": 0 00:15:09.847 }, 00:15:09.847 "claimed": true, 00:15:09.847 "claim_type": "exclusive_write", 00:15:09.847 "zoned": false, 00:15:09.847 "supported_io_types": { 00:15:09.847 "read": true, 00:15:09.847 "write": true, 00:15:09.847 "unmap": true, 00:15:09.847 "flush": true, 00:15:09.847 "reset": true, 00:15:09.847 "nvme_admin": false, 00:15:09.847 "nvme_io": false, 00:15:09.847 "nvme_io_md": false, 00:15:09.847 "write_zeroes": true, 00:15:09.847 "zcopy": true, 00:15:09.847 "get_zone_info": false, 00:15:09.847 "zone_management": false, 00:15:09.847 "zone_append": false, 00:15:09.847 "compare": false, 00:15:09.847 "compare_and_write": false, 00:15:09.847 "abort": true, 00:15:09.847 "seek_hole": false, 00:15:09.847 "seek_data": false, 00:15:09.847 "copy": true, 00:15:09.847 "nvme_iov_md": false 00:15:09.847 }, 00:15:09.847 "memory_domains": [ 00:15:09.847 { 00:15:09.847 "dma_device_id": "system", 00:15:09.847 "dma_device_type": 1 00:15:09.847 }, 00:15:09.847 { 00:15:09.847 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:09.847 "dma_device_type": 2 00:15:09.847 } 00:15:09.847 ], 00:15:09.847 "driver_specific": {} 00:15:09.847 } 00:15:09.847 ] 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:09.847 05:42:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.847 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:09.847 "name": "Existed_Raid", 00:15:09.847 "uuid": "ec813875-74cb-4a0b-bd07-3b1810ce1c63", 00:15:09.847 "strip_size_kb": 0, 00:15:09.847 "state": "configuring", 00:15:09.847 "raid_level": "raid1", 00:15:09.847 "superblock": true, 00:15:09.847 "num_base_bdevs": 2, 00:15:09.847 "num_base_bdevs_discovered": 1, 00:15:09.847 "num_base_bdevs_operational": 2, 00:15:09.847 "base_bdevs_list": [ 00:15:09.847 { 00:15:09.847 "name": "BaseBdev1", 00:15:09.847 "uuid": "faf6d498-d622-46fd-9cdd-aee5b2b35c23", 00:15:09.847 "is_configured": true, 00:15:09.847 "data_offset": 256, 00:15:09.847 "data_size": 7936 00:15:09.847 }, 00:15:09.847 { 00:15:09.847 "name": "BaseBdev2", 00:15:09.847 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:09.847 "is_configured": false, 00:15:09.847 "data_offset": 0, 00:15:09.847 "data_size": 0 00:15:09.847 } 00:15:09.847 ] 00:15:09.847 }' 00:15:09.847 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:09.847 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:10.107 [2024-12-07 05:42:43.406735] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:10.107 [2024-12-07 05:42:43.406847] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:10.107 [2024-12-07 05:42:43.418756] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:10.107 [2024-12-07 05:42:43.420622] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:10.107 [2024-12-07 05:42:43.420709] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:10.107 "name": "Existed_Raid", 00:15:10.107 "uuid": "cc55396d-3c9b-40f9-9aae-37f30b2fad94", 00:15:10.107 "strip_size_kb": 0, 00:15:10.107 "state": "configuring", 00:15:10.107 "raid_level": "raid1", 00:15:10.107 "superblock": true, 00:15:10.107 "num_base_bdevs": 2, 00:15:10.107 "num_base_bdevs_discovered": 1, 00:15:10.107 "num_base_bdevs_operational": 2, 00:15:10.107 "base_bdevs_list": [ 00:15:10.107 { 00:15:10.107 "name": "BaseBdev1", 00:15:10.107 "uuid": "faf6d498-d622-46fd-9cdd-aee5b2b35c23", 00:15:10.107 "is_configured": true, 00:15:10.107 "data_offset": 256, 00:15:10.107 "data_size": 7936 00:15:10.107 }, 00:15:10.107 { 00:15:10.107 "name": "BaseBdev2", 00:15:10.107 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:10.107 "is_configured": false, 00:15:10.107 "data_offset": 0, 00:15:10.107 "data_size": 0 00:15:10.107 } 00:15:10.107 ] 00:15:10.107 }' 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:10.107 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:10.677 [2024-12-07 05:42:43.816925] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:10.677 [2024-12-07 05:42:43.817232] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:10.677 [2024-12-07 05:42:43.817288] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:10.677 [2024-12-07 05:42:43.817581] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:10.677 BaseBdev2 00:15:10.677 [2024-12-07 05:42:43.817781] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:10.677 [2024-12-07 05:42:43.817797] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:15:10.677 [2024-12-07 05:42:43.817904] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.677 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:10.677 [ 00:15:10.677 { 00:15:10.677 "name": "BaseBdev2", 00:15:10.677 "aliases": [ 00:15:10.677 "288daea4-b3c1-45fd-8c25-0c9ea84bd80f" 00:15:10.677 ], 00:15:10.677 "product_name": "Malloc disk", 00:15:10.677 "block_size": 4096, 00:15:10.677 "num_blocks": 8192, 00:15:10.677 "uuid": "288daea4-b3c1-45fd-8c25-0c9ea84bd80f", 00:15:10.677 "assigned_rate_limits": { 00:15:10.677 "rw_ios_per_sec": 0, 00:15:10.677 "rw_mbytes_per_sec": 0, 00:15:10.677 "r_mbytes_per_sec": 0, 00:15:10.677 "w_mbytes_per_sec": 0 00:15:10.677 }, 00:15:10.677 "claimed": true, 00:15:10.677 "claim_type": "exclusive_write", 00:15:10.677 "zoned": false, 00:15:10.677 "supported_io_types": { 00:15:10.677 "read": true, 00:15:10.677 "write": true, 00:15:10.677 "unmap": true, 00:15:10.677 "flush": true, 00:15:10.677 "reset": true, 00:15:10.677 "nvme_admin": false, 00:15:10.677 "nvme_io": false, 00:15:10.677 "nvme_io_md": false, 00:15:10.678 "write_zeroes": true, 00:15:10.678 "zcopy": true, 00:15:10.678 "get_zone_info": false, 00:15:10.678 "zone_management": false, 00:15:10.678 "zone_append": false, 00:15:10.678 "compare": false, 00:15:10.678 "compare_and_write": false, 00:15:10.678 "abort": true, 00:15:10.678 "seek_hole": false, 00:15:10.678 "seek_data": false, 00:15:10.678 "copy": true, 00:15:10.678 "nvme_iov_md": false 00:15:10.678 }, 00:15:10.678 "memory_domains": [ 00:15:10.678 { 00:15:10.678 "dma_device_id": "system", 00:15:10.678 "dma_device_type": 1 00:15:10.678 }, 00:15:10.678 { 00:15:10.678 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:10.678 "dma_device_type": 2 00:15:10.678 } 00:15:10.678 ], 00:15:10.678 "driver_specific": {} 00:15:10.678 } 00:15:10.678 ] 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:10.678 "name": "Existed_Raid", 00:15:10.678 "uuid": "cc55396d-3c9b-40f9-9aae-37f30b2fad94", 00:15:10.678 "strip_size_kb": 0, 00:15:10.678 "state": "online", 00:15:10.678 "raid_level": "raid1", 00:15:10.678 "superblock": true, 00:15:10.678 "num_base_bdevs": 2, 00:15:10.678 "num_base_bdevs_discovered": 2, 00:15:10.678 "num_base_bdevs_operational": 2, 00:15:10.678 "base_bdevs_list": [ 00:15:10.678 { 00:15:10.678 "name": "BaseBdev1", 00:15:10.678 "uuid": "faf6d498-d622-46fd-9cdd-aee5b2b35c23", 00:15:10.678 "is_configured": true, 00:15:10.678 "data_offset": 256, 00:15:10.678 "data_size": 7936 00:15:10.678 }, 00:15:10.678 { 00:15:10.678 "name": "BaseBdev2", 00:15:10.678 "uuid": "288daea4-b3c1-45fd-8c25-0c9ea84bd80f", 00:15:10.678 "is_configured": true, 00:15:10.678 "data_offset": 256, 00:15:10.678 "data_size": 7936 00:15:10.678 } 00:15:10.678 ] 00:15:10.678 }' 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:10.678 05:42:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:11.248 [2024-12-07 05:42:44.320318] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:11.248 "name": "Existed_Raid", 00:15:11.248 "aliases": [ 00:15:11.248 "cc55396d-3c9b-40f9-9aae-37f30b2fad94" 00:15:11.248 ], 00:15:11.248 "product_name": "Raid Volume", 00:15:11.248 "block_size": 4096, 00:15:11.248 "num_blocks": 7936, 00:15:11.248 "uuid": "cc55396d-3c9b-40f9-9aae-37f30b2fad94", 00:15:11.248 "assigned_rate_limits": { 00:15:11.248 "rw_ios_per_sec": 0, 00:15:11.248 "rw_mbytes_per_sec": 0, 00:15:11.248 "r_mbytes_per_sec": 0, 00:15:11.248 "w_mbytes_per_sec": 0 00:15:11.248 }, 00:15:11.248 "claimed": false, 00:15:11.248 "zoned": false, 00:15:11.248 "supported_io_types": { 00:15:11.248 "read": true, 00:15:11.248 "write": true, 00:15:11.248 "unmap": false, 00:15:11.248 "flush": false, 00:15:11.248 "reset": true, 00:15:11.248 "nvme_admin": false, 00:15:11.248 "nvme_io": false, 00:15:11.248 "nvme_io_md": false, 00:15:11.248 "write_zeroes": true, 00:15:11.248 "zcopy": false, 00:15:11.248 "get_zone_info": false, 00:15:11.248 "zone_management": false, 00:15:11.248 "zone_append": false, 00:15:11.248 "compare": false, 00:15:11.248 "compare_and_write": false, 00:15:11.248 "abort": false, 00:15:11.248 "seek_hole": false, 00:15:11.248 "seek_data": false, 00:15:11.248 "copy": false, 00:15:11.248 "nvme_iov_md": false 00:15:11.248 }, 00:15:11.248 "memory_domains": [ 00:15:11.248 { 00:15:11.248 "dma_device_id": "system", 00:15:11.248 "dma_device_type": 1 00:15:11.248 }, 00:15:11.248 { 00:15:11.248 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:11.248 "dma_device_type": 2 00:15:11.248 }, 00:15:11.248 { 00:15:11.248 "dma_device_id": "system", 00:15:11.248 "dma_device_type": 1 00:15:11.248 }, 00:15:11.248 { 00:15:11.248 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:11.248 "dma_device_type": 2 00:15:11.248 } 00:15:11.248 ], 00:15:11.248 "driver_specific": { 00:15:11.248 "raid": { 00:15:11.248 "uuid": "cc55396d-3c9b-40f9-9aae-37f30b2fad94", 00:15:11.248 "strip_size_kb": 0, 00:15:11.248 "state": "online", 00:15:11.248 "raid_level": "raid1", 00:15:11.248 "superblock": true, 00:15:11.248 "num_base_bdevs": 2, 00:15:11.248 "num_base_bdevs_discovered": 2, 00:15:11.248 "num_base_bdevs_operational": 2, 00:15:11.248 "base_bdevs_list": [ 00:15:11.248 { 00:15:11.248 "name": "BaseBdev1", 00:15:11.248 "uuid": "faf6d498-d622-46fd-9cdd-aee5b2b35c23", 00:15:11.248 "is_configured": true, 00:15:11.248 "data_offset": 256, 00:15:11.248 "data_size": 7936 00:15:11.248 }, 00:15:11.248 { 00:15:11.248 "name": "BaseBdev2", 00:15:11.248 "uuid": "288daea4-b3c1-45fd-8c25-0c9ea84bd80f", 00:15:11.248 "is_configured": true, 00:15:11.248 "data_offset": 256, 00:15:11.248 "data_size": 7936 00:15:11.248 } 00:15:11.248 ] 00:15:11.248 } 00:15:11.248 } 00:15:11.248 }' 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:11.248 BaseBdev2' 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:11.248 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:11.249 [2024-12-07 05:42:44.503804] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:11.249 "name": "Existed_Raid", 00:15:11.249 "uuid": "cc55396d-3c9b-40f9-9aae-37f30b2fad94", 00:15:11.249 "strip_size_kb": 0, 00:15:11.249 "state": "online", 00:15:11.249 "raid_level": "raid1", 00:15:11.249 "superblock": true, 00:15:11.249 "num_base_bdevs": 2, 00:15:11.249 "num_base_bdevs_discovered": 1, 00:15:11.249 "num_base_bdevs_operational": 1, 00:15:11.249 "base_bdevs_list": [ 00:15:11.249 { 00:15:11.249 "name": null, 00:15:11.249 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:11.249 "is_configured": false, 00:15:11.249 "data_offset": 0, 00:15:11.249 "data_size": 7936 00:15:11.249 }, 00:15:11.249 { 00:15:11.249 "name": "BaseBdev2", 00:15:11.249 "uuid": "288daea4-b3c1-45fd-8c25-0c9ea84bd80f", 00:15:11.249 "is_configured": true, 00:15:11.249 "data_offset": 256, 00:15:11.249 "data_size": 7936 00:15:11.249 } 00:15:11.249 ] 00:15:11.249 }' 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:11.249 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:11.819 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:11.819 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:11.819 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.819 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:11.819 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.819 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:11.819 05:42:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:11.819 [2024-12-07 05:42:45.017973] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:11.819 [2024-12-07 05:42:45.018123] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:11.819 [2024-12-07 05:42:45.029716] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:11.819 [2024-12-07 05:42:45.029862] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:11.819 [2024-12-07 05:42:45.029917] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 95899 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 95899 ']' 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 95899 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 95899 00:15:11.819 killing process with pid 95899 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 95899' 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@973 -- # kill 95899 00:15:11.819 [2024-12-07 05:42:45.130407] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:11.819 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@978 -- # wait 95899 00:15:11.819 [2024-12-07 05:42:45.131369] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:12.078 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:15:12.078 00:15:12.078 real 0m3.773s 00:15:12.078 user 0m5.917s 00:15:12.078 sys 0m0.779s 00:15:12.078 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:12.078 05:42:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:12.078 ************************************ 00:15:12.078 END TEST raid_state_function_test_sb_4k 00:15:12.078 ************************************ 00:15:12.078 05:42:45 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:15:12.078 05:42:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:15:12.078 05:42:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:12.078 05:42:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:12.078 ************************************ 00:15:12.078 START TEST raid_superblock_test_4k 00:15:12.078 ************************************ 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=96136 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 96136 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@835 -- # '[' -z 96136 ']' 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:12.078 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:12.078 05:42:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:12.337 [2024-12-07 05:42:45.496853] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:15:12.337 [2024-12-07 05:42:45.497062] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96136 ] 00:15:12.337 [2024-12-07 05:42:45.651206] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:12.337 [2024-12-07 05:42:45.675372] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:12.597 [2024-12-07 05:42:45.717390] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:12.597 [2024-12-07 05:42:45.717509] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@868 -- # return 0 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.166 malloc1 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.166 [2024-12-07 05:42:46.340746] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:13.166 [2024-12-07 05:42:46.340925] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:13.166 [2024-12-07 05:42:46.340967] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:13.166 [2024-12-07 05:42:46.341017] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:13.166 [2024-12-07 05:42:46.343082] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:13.166 [2024-12-07 05:42:46.343177] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:13.166 pt1 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.166 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.167 malloc2 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.167 [2024-12-07 05:42:46.369043] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:13.167 [2024-12-07 05:42:46.369176] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:13.167 [2024-12-07 05:42:46.369214] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:13.167 [2024-12-07 05:42:46.369271] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:13.167 [2024-12-07 05:42:46.371331] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:13.167 [2024-12-07 05:42:46.371405] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:13.167 pt2 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.167 [2024-12-07 05:42:46.381070] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:13.167 [2024-12-07 05:42:46.382946] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:13.167 [2024-12-07 05:42:46.383139] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:13.167 [2024-12-07 05:42:46.383206] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:13.167 [2024-12-07 05:42:46.383505] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:13.167 [2024-12-07 05:42:46.383693] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:13.167 [2024-12-07 05:42:46.383741] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:13.167 [2024-12-07 05:42:46.383891] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:13.167 "name": "raid_bdev1", 00:15:13.167 "uuid": "2e2fba1e-4e89-42b1-b4b4-440e56c31edd", 00:15:13.167 "strip_size_kb": 0, 00:15:13.167 "state": "online", 00:15:13.167 "raid_level": "raid1", 00:15:13.167 "superblock": true, 00:15:13.167 "num_base_bdevs": 2, 00:15:13.167 "num_base_bdevs_discovered": 2, 00:15:13.167 "num_base_bdevs_operational": 2, 00:15:13.167 "base_bdevs_list": [ 00:15:13.167 { 00:15:13.167 "name": "pt1", 00:15:13.167 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:13.167 "is_configured": true, 00:15:13.167 "data_offset": 256, 00:15:13.167 "data_size": 7936 00:15:13.167 }, 00:15:13.167 { 00:15:13.167 "name": "pt2", 00:15:13.167 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:13.167 "is_configured": true, 00:15:13.167 "data_offset": 256, 00:15:13.167 "data_size": 7936 00:15:13.167 } 00:15:13.167 ] 00:15:13.167 }' 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:13.167 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.769 [2024-12-07 05:42:46.820559] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.769 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:13.769 "name": "raid_bdev1", 00:15:13.769 "aliases": [ 00:15:13.769 "2e2fba1e-4e89-42b1-b4b4-440e56c31edd" 00:15:13.769 ], 00:15:13.769 "product_name": "Raid Volume", 00:15:13.769 "block_size": 4096, 00:15:13.769 "num_blocks": 7936, 00:15:13.769 "uuid": "2e2fba1e-4e89-42b1-b4b4-440e56c31edd", 00:15:13.769 "assigned_rate_limits": { 00:15:13.769 "rw_ios_per_sec": 0, 00:15:13.769 "rw_mbytes_per_sec": 0, 00:15:13.769 "r_mbytes_per_sec": 0, 00:15:13.769 "w_mbytes_per_sec": 0 00:15:13.769 }, 00:15:13.769 "claimed": false, 00:15:13.769 "zoned": false, 00:15:13.769 "supported_io_types": { 00:15:13.769 "read": true, 00:15:13.769 "write": true, 00:15:13.769 "unmap": false, 00:15:13.769 "flush": false, 00:15:13.769 "reset": true, 00:15:13.769 "nvme_admin": false, 00:15:13.769 "nvme_io": false, 00:15:13.769 "nvme_io_md": false, 00:15:13.769 "write_zeroes": true, 00:15:13.769 "zcopy": false, 00:15:13.769 "get_zone_info": false, 00:15:13.769 "zone_management": false, 00:15:13.769 "zone_append": false, 00:15:13.769 "compare": false, 00:15:13.769 "compare_and_write": false, 00:15:13.769 "abort": false, 00:15:13.769 "seek_hole": false, 00:15:13.769 "seek_data": false, 00:15:13.769 "copy": false, 00:15:13.769 "nvme_iov_md": false 00:15:13.769 }, 00:15:13.769 "memory_domains": [ 00:15:13.769 { 00:15:13.769 "dma_device_id": "system", 00:15:13.769 "dma_device_type": 1 00:15:13.769 }, 00:15:13.769 { 00:15:13.769 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:13.769 "dma_device_type": 2 00:15:13.769 }, 00:15:13.769 { 00:15:13.769 "dma_device_id": "system", 00:15:13.769 "dma_device_type": 1 00:15:13.769 }, 00:15:13.769 { 00:15:13.769 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:13.769 "dma_device_type": 2 00:15:13.769 } 00:15:13.769 ], 00:15:13.769 "driver_specific": { 00:15:13.769 "raid": { 00:15:13.769 "uuid": "2e2fba1e-4e89-42b1-b4b4-440e56c31edd", 00:15:13.769 "strip_size_kb": 0, 00:15:13.769 "state": "online", 00:15:13.769 "raid_level": "raid1", 00:15:13.769 "superblock": true, 00:15:13.769 "num_base_bdevs": 2, 00:15:13.769 "num_base_bdevs_discovered": 2, 00:15:13.769 "num_base_bdevs_operational": 2, 00:15:13.769 "base_bdevs_list": [ 00:15:13.769 { 00:15:13.769 "name": "pt1", 00:15:13.769 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:13.769 "is_configured": true, 00:15:13.769 "data_offset": 256, 00:15:13.769 "data_size": 7936 00:15:13.769 }, 00:15:13.769 { 00:15:13.769 "name": "pt2", 00:15:13.770 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:13.770 "is_configured": true, 00:15:13.770 "data_offset": 256, 00:15:13.770 "data_size": 7936 00:15:13.770 } 00:15:13.770 ] 00:15:13.770 } 00:15:13.770 } 00:15:13.770 }' 00:15:13.770 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:13.770 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:13.770 pt2' 00:15:13.770 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:13.770 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:13.770 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:13.770 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:13.770 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.770 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.770 05:42:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:13.770 05:42:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.770 [2024-12-07 05:42:47.072041] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=2e2fba1e-4e89-42b1-b4b4-440e56c31edd 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z 2e2fba1e-4e89-42b1-b4b4-440e56c31edd ']' 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.770 [2024-12-07 05:42:47.119751] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:13.770 [2024-12-07 05:42:47.119821] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:13.770 [2024-12-07 05:42:47.119917] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:13.770 [2024-12-07 05:42:47.120015] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:13.770 [2024-12-07 05:42:47.120061] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:13.770 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # local es=0 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.044 [2024-12-07 05:42:47.247578] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:14.044 [2024-12-07 05:42:47.249462] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:14.044 [2024-12-07 05:42:47.249596] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:14.044 [2024-12-07 05:42:47.249717] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:14.044 [2024-12-07 05:42:47.249767] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:14.044 [2024-12-07 05:42:47.249813] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:15:14.044 request: 00:15:14.044 { 00:15:14.044 "name": "raid_bdev1", 00:15:14.044 "raid_level": "raid1", 00:15:14.044 "base_bdevs": [ 00:15:14.044 "malloc1", 00:15:14.044 "malloc2" 00:15:14.044 ], 00:15:14.044 "superblock": false, 00:15:14.044 "method": "bdev_raid_create", 00:15:14.044 "req_id": 1 00:15:14.044 } 00:15:14.044 Got JSON-RPC error response 00:15:14.044 response: 00:15:14.044 { 00:15:14.044 "code": -17, 00:15:14.044 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:14.044 } 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # es=1 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.044 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.044 [2024-12-07 05:42:47.311457] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:14.045 [2024-12-07 05:42:47.311559] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:14.045 [2024-12-07 05:42:47.311596] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:14.045 [2024-12-07 05:42:47.311635] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:14.045 [2024-12-07 05:42:47.313711] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:14.045 [2024-12-07 05:42:47.313778] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:14.045 [2024-12-07 05:42:47.313862] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:14.045 [2024-12-07 05:42:47.313910] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:14.045 pt1 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:14.045 "name": "raid_bdev1", 00:15:14.045 "uuid": "2e2fba1e-4e89-42b1-b4b4-440e56c31edd", 00:15:14.045 "strip_size_kb": 0, 00:15:14.045 "state": "configuring", 00:15:14.045 "raid_level": "raid1", 00:15:14.045 "superblock": true, 00:15:14.045 "num_base_bdevs": 2, 00:15:14.045 "num_base_bdevs_discovered": 1, 00:15:14.045 "num_base_bdevs_operational": 2, 00:15:14.045 "base_bdevs_list": [ 00:15:14.045 { 00:15:14.045 "name": "pt1", 00:15:14.045 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:14.045 "is_configured": true, 00:15:14.045 "data_offset": 256, 00:15:14.045 "data_size": 7936 00:15:14.045 }, 00:15:14.045 { 00:15:14.045 "name": null, 00:15:14.045 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:14.045 "is_configured": false, 00:15:14.045 "data_offset": 256, 00:15:14.045 "data_size": 7936 00:15:14.045 } 00:15:14.045 ] 00:15:14.045 }' 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:14.045 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.615 [2024-12-07 05:42:47.746725] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:14.615 [2024-12-07 05:42:47.746838] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:14.615 [2024-12-07 05:42:47.746874] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:14.615 [2024-12-07 05:42:47.746901] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:14.615 [2024-12-07 05:42:47.747250] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:14.615 [2024-12-07 05:42:47.747307] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:14.615 [2024-12-07 05:42:47.747389] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:14.615 [2024-12-07 05:42:47.747437] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:14.615 [2024-12-07 05:42:47.747549] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:14.615 [2024-12-07 05:42:47.747588] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:14.615 [2024-12-07 05:42:47.747849] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:14.615 [2024-12-07 05:42:47.747995] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:14.615 [2024-12-07 05:42:47.748040] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:15:14.615 [2024-12-07 05:42:47.748173] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:14.615 pt2 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:14.615 "name": "raid_bdev1", 00:15:14.615 "uuid": "2e2fba1e-4e89-42b1-b4b4-440e56c31edd", 00:15:14.615 "strip_size_kb": 0, 00:15:14.615 "state": "online", 00:15:14.615 "raid_level": "raid1", 00:15:14.615 "superblock": true, 00:15:14.615 "num_base_bdevs": 2, 00:15:14.615 "num_base_bdevs_discovered": 2, 00:15:14.615 "num_base_bdevs_operational": 2, 00:15:14.615 "base_bdevs_list": [ 00:15:14.615 { 00:15:14.615 "name": "pt1", 00:15:14.615 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:14.615 "is_configured": true, 00:15:14.615 "data_offset": 256, 00:15:14.615 "data_size": 7936 00:15:14.615 }, 00:15:14.615 { 00:15:14.615 "name": "pt2", 00:15:14.615 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:14.615 "is_configured": true, 00:15:14.615 "data_offset": 256, 00:15:14.615 "data_size": 7936 00:15:14.615 } 00:15:14.615 ] 00:15:14.615 }' 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:14.615 05:42:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:14.892 [2024-12-07 05:42:48.110350] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:14.892 "name": "raid_bdev1", 00:15:14.892 "aliases": [ 00:15:14.892 "2e2fba1e-4e89-42b1-b4b4-440e56c31edd" 00:15:14.892 ], 00:15:14.892 "product_name": "Raid Volume", 00:15:14.892 "block_size": 4096, 00:15:14.892 "num_blocks": 7936, 00:15:14.892 "uuid": "2e2fba1e-4e89-42b1-b4b4-440e56c31edd", 00:15:14.892 "assigned_rate_limits": { 00:15:14.892 "rw_ios_per_sec": 0, 00:15:14.892 "rw_mbytes_per_sec": 0, 00:15:14.892 "r_mbytes_per_sec": 0, 00:15:14.892 "w_mbytes_per_sec": 0 00:15:14.892 }, 00:15:14.892 "claimed": false, 00:15:14.892 "zoned": false, 00:15:14.892 "supported_io_types": { 00:15:14.892 "read": true, 00:15:14.892 "write": true, 00:15:14.892 "unmap": false, 00:15:14.892 "flush": false, 00:15:14.892 "reset": true, 00:15:14.892 "nvme_admin": false, 00:15:14.892 "nvme_io": false, 00:15:14.892 "nvme_io_md": false, 00:15:14.892 "write_zeroes": true, 00:15:14.892 "zcopy": false, 00:15:14.892 "get_zone_info": false, 00:15:14.892 "zone_management": false, 00:15:14.892 "zone_append": false, 00:15:14.892 "compare": false, 00:15:14.892 "compare_and_write": false, 00:15:14.892 "abort": false, 00:15:14.892 "seek_hole": false, 00:15:14.892 "seek_data": false, 00:15:14.892 "copy": false, 00:15:14.892 "nvme_iov_md": false 00:15:14.892 }, 00:15:14.892 "memory_domains": [ 00:15:14.892 { 00:15:14.892 "dma_device_id": "system", 00:15:14.892 "dma_device_type": 1 00:15:14.892 }, 00:15:14.892 { 00:15:14.892 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:14.892 "dma_device_type": 2 00:15:14.892 }, 00:15:14.892 { 00:15:14.892 "dma_device_id": "system", 00:15:14.892 "dma_device_type": 1 00:15:14.892 }, 00:15:14.892 { 00:15:14.892 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:14.892 "dma_device_type": 2 00:15:14.892 } 00:15:14.892 ], 00:15:14.892 "driver_specific": { 00:15:14.892 "raid": { 00:15:14.892 "uuid": "2e2fba1e-4e89-42b1-b4b4-440e56c31edd", 00:15:14.892 "strip_size_kb": 0, 00:15:14.892 "state": "online", 00:15:14.892 "raid_level": "raid1", 00:15:14.892 "superblock": true, 00:15:14.892 "num_base_bdevs": 2, 00:15:14.892 "num_base_bdevs_discovered": 2, 00:15:14.892 "num_base_bdevs_operational": 2, 00:15:14.892 "base_bdevs_list": [ 00:15:14.892 { 00:15:14.892 "name": "pt1", 00:15:14.892 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:14.892 "is_configured": true, 00:15:14.892 "data_offset": 256, 00:15:14.892 "data_size": 7936 00:15:14.892 }, 00:15:14.892 { 00:15:14.892 "name": "pt2", 00:15:14.892 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:14.892 "is_configured": true, 00:15:14.892 "data_offset": 256, 00:15:14.892 "data_size": 7936 00:15:14.892 } 00:15:14.892 ] 00:15:14.892 } 00:15:14.892 } 00:15:14.892 }' 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:14.892 pt2' 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:14.892 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:14.893 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:14.893 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.893 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.153 [2024-12-07 05:42:48.333953] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' 2e2fba1e-4e89-42b1-b4b4-440e56c31edd '!=' 2e2fba1e-4e89-42b1-b4b4-440e56c31edd ']' 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.153 [2024-12-07 05:42:48.381675] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:15.153 "name": "raid_bdev1", 00:15:15.153 "uuid": "2e2fba1e-4e89-42b1-b4b4-440e56c31edd", 00:15:15.153 "strip_size_kb": 0, 00:15:15.153 "state": "online", 00:15:15.153 "raid_level": "raid1", 00:15:15.153 "superblock": true, 00:15:15.153 "num_base_bdevs": 2, 00:15:15.153 "num_base_bdevs_discovered": 1, 00:15:15.153 "num_base_bdevs_operational": 1, 00:15:15.153 "base_bdevs_list": [ 00:15:15.153 { 00:15:15.153 "name": null, 00:15:15.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:15.153 "is_configured": false, 00:15:15.153 "data_offset": 0, 00:15:15.153 "data_size": 7936 00:15:15.153 }, 00:15:15.153 { 00:15:15.153 "name": "pt2", 00:15:15.153 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:15.153 "is_configured": true, 00:15:15.153 "data_offset": 256, 00:15:15.153 "data_size": 7936 00:15:15.153 } 00:15:15.153 ] 00:15:15.153 }' 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:15.153 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.722 [2024-12-07 05:42:48.796885] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:15.722 [2024-12-07 05:42:48.796963] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:15.722 [2024-12-07 05:42:48.797068] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:15.722 [2024-12-07 05:42:48.797130] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:15.722 [2024-12-07 05:42:48.797187] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.722 [2024-12-07 05:42:48.872754] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:15.722 [2024-12-07 05:42:48.872846] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:15.722 [2024-12-07 05:42:48.872879] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:15:15.722 [2024-12-07 05:42:48.872905] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:15.722 [2024-12-07 05:42:48.875126] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:15.722 [2024-12-07 05:42:48.875216] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:15.722 [2024-12-07 05:42:48.875304] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:15.722 [2024-12-07 05:42:48.875354] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:15.722 [2024-12-07 05:42:48.875464] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:15:15.722 [2024-12-07 05:42:48.875502] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:15.722 [2024-12-07 05:42:48.875756] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:15.722 [2024-12-07 05:42:48.875904] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:15:15.722 [2024-12-07 05:42:48.875918] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:15:15.722 [2024-12-07 05:42:48.876017] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:15.722 pt2 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:15.722 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:15.723 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:15.723 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:15.723 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.723 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.723 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.723 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.723 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.723 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:15.723 "name": "raid_bdev1", 00:15:15.723 "uuid": "2e2fba1e-4e89-42b1-b4b4-440e56c31edd", 00:15:15.723 "strip_size_kb": 0, 00:15:15.723 "state": "online", 00:15:15.723 "raid_level": "raid1", 00:15:15.723 "superblock": true, 00:15:15.723 "num_base_bdevs": 2, 00:15:15.723 "num_base_bdevs_discovered": 1, 00:15:15.723 "num_base_bdevs_operational": 1, 00:15:15.723 "base_bdevs_list": [ 00:15:15.723 { 00:15:15.723 "name": null, 00:15:15.723 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:15.723 "is_configured": false, 00:15:15.723 "data_offset": 256, 00:15:15.723 "data_size": 7936 00:15:15.723 }, 00:15:15.723 { 00:15:15.723 "name": "pt2", 00:15:15.723 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:15.723 "is_configured": true, 00:15:15.723 "data_offset": 256, 00:15:15.723 "data_size": 7936 00:15:15.723 } 00:15:15.723 ] 00:15:15.723 }' 00:15:15.723 05:42:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:15.723 05:42:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.984 [2024-12-07 05:42:49.264137] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:15.984 [2024-12-07 05:42:49.264221] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:15.984 [2024-12-07 05:42:49.264324] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:15.984 [2024-12-07 05:42:49.264386] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:15.984 [2024-12-07 05:42:49.264456] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.984 [2024-12-07 05:42:49.324004] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:15.984 [2024-12-07 05:42:49.324126] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:15.984 [2024-12-07 05:42:49.324161] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:15:15.984 [2024-12-07 05:42:49.324194] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:15.984 [2024-12-07 05:42:49.326316] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:15.984 [2024-12-07 05:42:49.326411] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:15.984 [2024-12-07 05:42:49.326505] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:15.984 [2024-12-07 05:42:49.326572] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:15.984 [2024-12-07 05:42:49.326712] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:15.984 [2024-12-07 05:42:49.326775] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:15.984 [2024-12-07 05:42:49.326812] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:15:15.984 [2024-12-07 05:42:49.326889] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:15.984 [2024-12-07 05:42:49.326999] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:15:15.984 [2024-12-07 05:42:49.327041] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:15.984 [2024-12-07 05:42:49.327276] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:15.984 [2024-12-07 05:42:49.327423] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:15:15.984 [2024-12-07 05:42:49.327462] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:15:15.984 [2024-12-07 05:42:49.327606] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:15.984 pt1 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.984 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.244 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:16.244 "name": "raid_bdev1", 00:15:16.244 "uuid": "2e2fba1e-4e89-42b1-b4b4-440e56c31edd", 00:15:16.244 "strip_size_kb": 0, 00:15:16.244 "state": "online", 00:15:16.244 "raid_level": "raid1", 00:15:16.244 "superblock": true, 00:15:16.244 "num_base_bdevs": 2, 00:15:16.244 "num_base_bdevs_discovered": 1, 00:15:16.244 "num_base_bdevs_operational": 1, 00:15:16.244 "base_bdevs_list": [ 00:15:16.244 { 00:15:16.244 "name": null, 00:15:16.244 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:16.244 "is_configured": false, 00:15:16.244 "data_offset": 256, 00:15:16.244 "data_size": 7936 00:15:16.244 }, 00:15:16.244 { 00:15:16.244 "name": "pt2", 00:15:16.244 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:16.244 "is_configured": true, 00:15:16.244 "data_offset": 256, 00:15:16.244 "data_size": 7936 00:15:16.244 } 00:15:16.244 ] 00:15:16.244 }' 00:15:16.244 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:16.244 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:16.504 [2024-12-07 05:42:49.791403] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' 2e2fba1e-4e89-42b1-b4b4-440e56c31edd '!=' 2e2fba1e-4e89-42b1-b4b4-440e56c31edd ']' 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 96136 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # '[' -z 96136 ']' 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@958 -- # kill -0 96136 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # uname 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96136 00:15:16.504 killing process with pid 96136 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96136' 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@973 -- # kill 96136 00:15:16.504 [2024-12-07 05:42:49.847535] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:16.504 [2024-12-07 05:42:49.847600] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:16.504 [2024-12-07 05:42:49.847661] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:16.504 [2024-12-07 05:42:49.847671] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:15:16.504 05:42:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@978 -- # wait 96136 00:15:16.763 [2024-12-07 05:42:49.869963] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:16.763 05:42:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:15:16.763 00:15:16.763 real 0m4.667s 00:15:16.763 user 0m7.628s 00:15:16.763 sys 0m0.970s 00:15:16.763 05:42:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:16.763 05:42:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:16.763 ************************************ 00:15:16.763 END TEST raid_superblock_test_4k 00:15:16.763 ************************************ 00:15:17.023 05:42:50 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:15:17.023 05:42:50 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:15:17.023 05:42:50 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:17.023 05:42:50 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:17.023 05:42:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:17.023 ************************************ 00:15:17.023 START TEST raid_rebuild_test_sb_4k 00:15:17.023 ************************************ 00:15:17.023 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:15:17.023 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:15:17.023 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:15:17.023 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:17.023 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:17.023 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:17.023 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:17.023 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=96446 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 96446 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 96446 ']' 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:17.024 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:17.024 05:42:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:17.024 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:17.024 Zero copy mechanism will not be used. 00:15:17.024 [2024-12-07 05:42:50.248983] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:15:17.024 [2024-12-07 05:42:50.249106] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96446 ] 00:15:17.284 [2024-12-07 05:42:50.402286] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:17.284 [2024-12-07 05:42:50.426514] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:17.284 [2024-12-07 05:42:50.468271] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:17.284 [2024-12-07 05:42:50.468307] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:17.855 BaseBdev1_malloc 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:17.855 [2024-12-07 05:42:51.087088] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:17.855 [2024-12-07 05:42:51.087265] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:17.855 [2024-12-07 05:42:51.087320] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:17.855 [2024-12-07 05:42:51.087390] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:17.855 [2024-12-07 05:42:51.089489] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:17.855 [2024-12-07 05:42:51.089564] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:17.855 BaseBdev1 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:17.855 BaseBdev2_malloc 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:17.855 [2024-12-07 05:42:51.115440] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:17.855 [2024-12-07 05:42:51.115558] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:17.855 [2024-12-07 05:42:51.115599] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:17.855 [2024-12-07 05:42:51.115635] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:17.855 [2024-12-07 05:42:51.117696] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:17.855 [2024-12-07 05:42:51.117771] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:17.855 BaseBdev2 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:17.855 spare_malloc 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:17.855 spare_delay 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:17.855 [2024-12-07 05:42:51.155668] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:17.855 [2024-12-07 05:42:51.155787] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:17.855 [2024-12-07 05:42:51.155822] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:17.855 [2024-12-07 05:42:51.155849] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:17.855 [2024-12-07 05:42:51.157927] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:17.855 [2024-12-07 05:42:51.157993] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:17.855 spare 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.855 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:17.855 [2024-12-07 05:42:51.167690] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:17.855 [2024-12-07 05:42:51.169556] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:17.855 [2024-12-07 05:42:51.169784] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:17.855 [2024-12-07 05:42:51.169802] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:17.855 [2024-12-07 05:42:51.170074] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:17.855 [2024-12-07 05:42:51.170212] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:17.856 [2024-12-07 05:42:51.170224] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:17.856 [2024-12-07 05:42:51.170332] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:17.856 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.116 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:18.116 "name": "raid_bdev1", 00:15:18.116 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:18.116 "strip_size_kb": 0, 00:15:18.116 "state": "online", 00:15:18.116 "raid_level": "raid1", 00:15:18.116 "superblock": true, 00:15:18.116 "num_base_bdevs": 2, 00:15:18.116 "num_base_bdevs_discovered": 2, 00:15:18.116 "num_base_bdevs_operational": 2, 00:15:18.116 "base_bdevs_list": [ 00:15:18.116 { 00:15:18.116 "name": "BaseBdev1", 00:15:18.116 "uuid": "4e7ca387-7cd9-5468-be13-32c0a056e32e", 00:15:18.116 "is_configured": true, 00:15:18.116 "data_offset": 256, 00:15:18.116 "data_size": 7936 00:15:18.116 }, 00:15:18.116 { 00:15:18.116 "name": "BaseBdev2", 00:15:18.116 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:18.116 "is_configured": true, 00:15:18.116 "data_offset": 256, 00:15:18.116 "data_size": 7936 00:15:18.116 } 00:15:18.116 ] 00:15:18.116 }' 00:15:18.116 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:18.116 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:18.376 [2024-12-07 05:42:51.607149] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:18.376 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:18.636 [2024-12-07 05:42:51.882578] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:18.636 /dev/nbd0 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:18.636 1+0 records in 00:15:18.636 1+0 records out 00:15:18.636 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000563065 s, 7.3 MB/s 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:15:18.636 05:42:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:15:19.575 7936+0 records in 00:15:19.575 7936+0 records out 00:15:19.575 32505856 bytes (33 MB, 31 MiB) copied, 0.631525 s, 51.5 MB/s 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:19.575 [2024-12-07 05:42:52.782611] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:19.575 [2024-12-07 05:42:52.798950] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:19.575 "name": "raid_bdev1", 00:15:19.575 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:19.575 "strip_size_kb": 0, 00:15:19.575 "state": "online", 00:15:19.575 "raid_level": "raid1", 00:15:19.575 "superblock": true, 00:15:19.575 "num_base_bdevs": 2, 00:15:19.575 "num_base_bdevs_discovered": 1, 00:15:19.575 "num_base_bdevs_operational": 1, 00:15:19.575 "base_bdevs_list": [ 00:15:19.575 { 00:15:19.575 "name": null, 00:15:19.575 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:19.575 "is_configured": false, 00:15:19.575 "data_offset": 0, 00:15:19.575 "data_size": 7936 00:15:19.575 }, 00:15:19.575 { 00:15:19.575 "name": "BaseBdev2", 00:15:19.575 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:19.575 "is_configured": true, 00:15:19.575 "data_offset": 256, 00:15:19.575 "data_size": 7936 00:15:19.575 } 00:15:19.575 ] 00:15:19.575 }' 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:19.575 05:42:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.144 05:42:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:20.144 05:42:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:20.144 05:42:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.144 [2024-12-07 05:42:53.262193] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:20.144 [2024-12-07 05:42:53.267084] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:15:20.144 05:42:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:20.144 05:42:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:20.144 [2024-12-07 05:42:53.268936] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:21.083 "name": "raid_bdev1", 00:15:21.083 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:21.083 "strip_size_kb": 0, 00:15:21.083 "state": "online", 00:15:21.083 "raid_level": "raid1", 00:15:21.083 "superblock": true, 00:15:21.083 "num_base_bdevs": 2, 00:15:21.083 "num_base_bdevs_discovered": 2, 00:15:21.083 "num_base_bdevs_operational": 2, 00:15:21.083 "process": { 00:15:21.083 "type": "rebuild", 00:15:21.083 "target": "spare", 00:15:21.083 "progress": { 00:15:21.083 "blocks": 2560, 00:15:21.083 "percent": 32 00:15:21.083 } 00:15:21.083 }, 00:15:21.083 "base_bdevs_list": [ 00:15:21.083 { 00:15:21.083 "name": "spare", 00:15:21.083 "uuid": "abb70d3f-1d43-52d1-bee5-e651b9148266", 00:15:21.083 "is_configured": true, 00:15:21.083 "data_offset": 256, 00:15:21.083 "data_size": 7936 00:15:21.083 }, 00:15:21.083 { 00:15:21.083 "name": "BaseBdev2", 00:15:21.083 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:21.083 "is_configured": true, 00:15:21.083 "data_offset": 256, 00:15:21.083 "data_size": 7936 00:15:21.083 } 00:15:21.083 ] 00:15:21.083 }' 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.083 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.083 [2024-12-07 05:42:54.429023] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:21.343 [2024-12-07 05:42:54.473427] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:21.343 [2024-12-07 05:42:54.473485] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:21.343 [2024-12-07 05:42:54.473519] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:21.343 [2024-12-07 05:42:54.473526] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:21.343 "name": "raid_bdev1", 00:15:21.343 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:21.343 "strip_size_kb": 0, 00:15:21.343 "state": "online", 00:15:21.343 "raid_level": "raid1", 00:15:21.343 "superblock": true, 00:15:21.343 "num_base_bdevs": 2, 00:15:21.343 "num_base_bdevs_discovered": 1, 00:15:21.343 "num_base_bdevs_operational": 1, 00:15:21.343 "base_bdevs_list": [ 00:15:21.343 { 00:15:21.343 "name": null, 00:15:21.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:21.343 "is_configured": false, 00:15:21.343 "data_offset": 0, 00:15:21.343 "data_size": 7936 00:15:21.343 }, 00:15:21.343 { 00:15:21.343 "name": "BaseBdev2", 00:15:21.343 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:21.343 "is_configured": true, 00:15:21.343 "data_offset": 256, 00:15:21.343 "data_size": 7936 00:15:21.343 } 00:15:21.343 ] 00:15:21.343 }' 00:15:21.343 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:21.344 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.603 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:21.603 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:21.603 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:21.603 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:21.603 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:21.603 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.603 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.603 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.603 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.863 05:42:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.864 05:42:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:21.864 "name": "raid_bdev1", 00:15:21.864 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:21.864 "strip_size_kb": 0, 00:15:21.864 "state": "online", 00:15:21.864 "raid_level": "raid1", 00:15:21.864 "superblock": true, 00:15:21.864 "num_base_bdevs": 2, 00:15:21.864 "num_base_bdevs_discovered": 1, 00:15:21.864 "num_base_bdevs_operational": 1, 00:15:21.864 "base_bdevs_list": [ 00:15:21.864 { 00:15:21.864 "name": null, 00:15:21.864 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:21.864 "is_configured": false, 00:15:21.864 "data_offset": 0, 00:15:21.864 "data_size": 7936 00:15:21.864 }, 00:15:21.864 { 00:15:21.864 "name": "BaseBdev2", 00:15:21.864 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:21.864 "is_configured": true, 00:15:21.864 "data_offset": 256, 00:15:21.864 "data_size": 7936 00:15:21.864 } 00:15:21.864 ] 00:15:21.864 }' 00:15:21.864 05:42:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:21.864 05:42:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:21.864 05:42:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:21.864 05:42:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:21.864 05:42:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:21.864 05:42:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.864 05:42:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.864 [2024-12-07 05:42:55.085216] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:21.864 [2024-12-07 05:42:55.090005] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:15:21.864 05:42:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.864 05:42:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:21.864 [2024-12-07 05:42:55.091873] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:22.799 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:22.799 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:22.799 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:22.799 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:22.800 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:22.800 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.800 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:22.800 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.800 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.800 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.800 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:22.800 "name": "raid_bdev1", 00:15:22.800 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:22.800 "strip_size_kb": 0, 00:15:22.800 "state": "online", 00:15:22.800 "raid_level": "raid1", 00:15:22.800 "superblock": true, 00:15:22.800 "num_base_bdevs": 2, 00:15:22.800 "num_base_bdevs_discovered": 2, 00:15:22.800 "num_base_bdevs_operational": 2, 00:15:22.800 "process": { 00:15:22.800 "type": "rebuild", 00:15:22.800 "target": "spare", 00:15:22.800 "progress": { 00:15:22.800 "blocks": 2560, 00:15:22.800 "percent": 32 00:15:22.800 } 00:15:22.800 }, 00:15:22.800 "base_bdevs_list": [ 00:15:22.800 { 00:15:22.800 "name": "spare", 00:15:22.800 "uuid": "abb70d3f-1d43-52d1-bee5-e651b9148266", 00:15:22.800 "is_configured": true, 00:15:22.800 "data_offset": 256, 00:15:22.800 "data_size": 7936 00:15:22.800 }, 00:15:22.800 { 00:15:22.800 "name": "BaseBdev2", 00:15:22.800 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:22.800 "is_configured": true, 00:15:22.800 "data_offset": 256, 00:15:22.800 "data_size": 7936 00:15:22.800 } 00:15:22.800 ] 00:15:22.800 }' 00:15:22.800 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:23.059 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=553 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:23.059 "name": "raid_bdev1", 00:15:23.059 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:23.059 "strip_size_kb": 0, 00:15:23.059 "state": "online", 00:15:23.059 "raid_level": "raid1", 00:15:23.059 "superblock": true, 00:15:23.059 "num_base_bdevs": 2, 00:15:23.059 "num_base_bdevs_discovered": 2, 00:15:23.059 "num_base_bdevs_operational": 2, 00:15:23.059 "process": { 00:15:23.059 "type": "rebuild", 00:15:23.059 "target": "spare", 00:15:23.059 "progress": { 00:15:23.059 "blocks": 2816, 00:15:23.059 "percent": 35 00:15:23.059 } 00:15:23.059 }, 00:15:23.059 "base_bdevs_list": [ 00:15:23.059 { 00:15:23.059 "name": "spare", 00:15:23.059 "uuid": "abb70d3f-1d43-52d1-bee5-e651b9148266", 00:15:23.059 "is_configured": true, 00:15:23.059 "data_offset": 256, 00:15:23.059 "data_size": 7936 00:15:23.059 }, 00:15:23.059 { 00:15:23.059 "name": "BaseBdev2", 00:15:23.059 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:23.059 "is_configured": true, 00:15:23.059 "data_offset": 256, 00:15:23.059 "data_size": 7936 00:15:23.059 } 00:15:23.059 ] 00:15:23.059 }' 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:23.059 05:42:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:24.434 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:24.435 "name": "raid_bdev1", 00:15:24.435 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:24.435 "strip_size_kb": 0, 00:15:24.435 "state": "online", 00:15:24.435 "raid_level": "raid1", 00:15:24.435 "superblock": true, 00:15:24.435 "num_base_bdevs": 2, 00:15:24.435 "num_base_bdevs_discovered": 2, 00:15:24.435 "num_base_bdevs_operational": 2, 00:15:24.435 "process": { 00:15:24.435 "type": "rebuild", 00:15:24.435 "target": "spare", 00:15:24.435 "progress": { 00:15:24.435 "blocks": 5632, 00:15:24.435 "percent": 70 00:15:24.435 } 00:15:24.435 }, 00:15:24.435 "base_bdevs_list": [ 00:15:24.435 { 00:15:24.435 "name": "spare", 00:15:24.435 "uuid": "abb70d3f-1d43-52d1-bee5-e651b9148266", 00:15:24.435 "is_configured": true, 00:15:24.435 "data_offset": 256, 00:15:24.435 "data_size": 7936 00:15:24.435 }, 00:15:24.435 { 00:15:24.435 "name": "BaseBdev2", 00:15:24.435 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:24.435 "is_configured": true, 00:15:24.435 "data_offset": 256, 00:15:24.435 "data_size": 7936 00:15:24.435 } 00:15:24.435 ] 00:15:24.435 }' 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:24.435 05:42:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:25.002 [2024-12-07 05:42:58.202007] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:25.002 [2024-12-07 05:42:58.202101] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:25.002 [2024-12-07 05:42:58.202214] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:25.260 "name": "raid_bdev1", 00:15:25.260 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:25.260 "strip_size_kb": 0, 00:15:25.260 "state": "online", 00:15:25.260 "raid_level": "raid1", 00:15:25.260 "superblock": true, 00:15:25.260 "num_base_bdevs": 2, 00:15:25.260 "num_base_bdevs_discovered": 2, 00:15:25.260 "num_base_bdevs_operational": 2, 00:15:25.260 "base_bdevs_list": [ 00:15:25.260 { 00:15:25.260 "name": "spare", 00:15:25.260 "uuid": "abb70d3f-1d43-52d1-bee5-e651b9148266", 00:15:25.260 "is_configured": true, 00:15:25.260 "data_offset": 256, 00:15:25.260 "data_size": 7936 00:15:25.260 }, 00:15:25.260 { 00:15:25.260 "name": "BaseBdev2", 00:15:25.260 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:25.260 "is_configured": true, 00:15:25.260 "data_offset": 256, 00:15:25.260 "data_size": 7936 00:15:25.260 } 00:15:25.260 ] 00:15:25.260 }' 00:15:25.260 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:25.518 "name": "raid_bdev1", 00:15:25.518 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:25.518 "strip_size_kb": 0, 00:15:25.518 "state": "online", 00:15:25.518 "raid_level": "raid1", 00:15:25.518 "superblock": true, 00:15:25.518 "num_base_bdevs": 2, 00:15:25.518 "num_base_bdevs_discovered": 2, 00:15:25.518 "num_base_bdevs_operational": 2, 00:15:25.518 "base_bdevs_list": [ 00:15:25.518 { 00:15:25.518 "name": "spare", 00:15:25.518 "uuid": "abb70d3f-1d43-52d1-bee5-e651b9148266", 00:15:25.518 "is_configured": true, 00:15:25.518 "data_offset": 256, 00:15:25.518 "data_size": 7936 00:15:25.518 }, 00:15:25.518 { 00:15:25.518 "name": "BaseBdev2", 00:15:25.518 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:25.518 "is_configured": true, 00:15:25.518 "data_offset": 256, 00:15:25.518 "data_size": 7936 00:15:25.518 } 00:15:25.518 ] 00:15:25.518 }' 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:25.518 "name": "raid_bdev1", 00:15:25.518 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:25.518 "strip_size_kb": 0, 00:15:25.518 "state": "online", 00:15:25.518 "raid_level": "raid1", 00:15:25.518 "superblock": true, 00:15:25.518 "num_base_bdevs": 2, 00:15:25.518 "num_base_bdevs_discovered": 2, 00:15:25.518 "num_base_bdevs_operational": 2, 00:15:25.518 "base_bdevs_list": [ 00:15:25.518 { 00:15:25.518 "name": "spare", 00:15:25.518 "uuid": "abb70d3f-1d43-52d1-bee5-e651b9148266", 00:15:25.518 "is_configured": true, 00:15:25.518 "data_offset": 256, 00:15:25.518 "data_size": 7936 00:15:25.518 }, 00:15:25.518 { 00:15:25.518 "name": "BaseBdev2", 00:15:25.518 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:25.518 "is_configured": true, 00:15:25.518 "data_offset": 256, 00:15:25.518 "data_size": 7936 00:15:25.518 } 00:15:25.518 ] 00:15:25.518 }' 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:25.518 05:42:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.085 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:26.085 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:26.085 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.085 [2024-12-07 05:42:59.257052] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:26.085 [2024-12-07 05:42:59.257085] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:26.085 [2024-12-07 05:42:59.257179] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:26.085 [2024-12-07 05:42:59.257261] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:26.085 [2024-12-07 05:42:59.257286] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:26.085 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:26.085 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:26.085 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:26.085 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.085 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:15:26.085 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:26.085 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:26.085 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:26.086 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:26.086 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:26.086 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:26.086 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:26.086 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:26.086 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:26.086 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:26.086 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:15:26.086 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:26.086 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:26.086 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:26.345 /dev/nbd0 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:26.345 1+0 records in 00:15:26.345 1+0 records out 00:15:26.345 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000317254 s, 12.9 MB/s 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:26.345 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:26.605 /dev/nbd1 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:26.605 1+0 records in 00:15:26.605 1+0 records out 00:15:26.605 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000440613 s, 9.3 MB/s 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:26.605 05:42:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:26.865 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:26.865 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:26.865 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:26.865 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:26.865 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:26.865 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:26.865 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:26.865 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:26.865 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:26.865 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.125 [2024-12-07 05:43:00.312737] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:27.125 [2024-12-07 05:43:00.312845] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:27.125 [2024-12-07 05:43:00.312881] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:27.125 [2024-12-07 05:43:00.312913] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:27.125 [2024-12-07 05:43:00.315051] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:27.125 [2024-12-07 05:43:00.315131] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:27.125 [2024-12-07 05:43:00.315256] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:27.125 [2024-12-07 05:43:00.315330] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:27.125 [2024-12-07 05:43:00.315490] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:27.125 spare 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.125 [2024-12-07 05:43:00.415429] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:27.125 [2024-12-07 05:43:00.415496] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:27.125 [2024-12-07 05:43:00.415771] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:15:27.125 [2024-12-07 05:43:00.415968] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:27.125 [2024-12-07 05:43:00.416016] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:27.125 [2024-12-07 05:43:00.416176] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:27.125 "name": "raid_bdev1", 00:15:27.125 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:27.125 "strip_size_kb": 0, 00:15:27.125 "state": "online", 00:15:27.125 "raid_level": "raid1", 00:15:27.125 "superblock": true, 00:15:27.125 "num_base_bdevs": 2, 00:15:27.125 "num_base_bdevs_discovered": 2, 00:15:27.125 "num_base_bdevs_operational": 2, 00:15:27.125 "base_bdevs_list": [ 00:15:27.125 { 00:15:27.125 "name": "spare", 00:15:27.125 "uuid": "abb70d3f-1d43-52d1-bee5-e651b9148266", 00:15:27.125 "is_configured": true, 00:15:27.125 "data_offset": 256, 00:15:27.125 "data_size": 7936 00:15:27.125 }, 00:15:27.125 { 00:15:27.125 "name": "BaseBdev2", 00:15:27.125 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:27.125 "is_configured": true, 00:15:27.125 "data_offset": 256, 00:15:27.125 "data_size": 7936 00:15:27.125 } 00:15:27.125 ] 00:15:27.125 }' 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:27.125 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:27.695 "name": "raid_bdev1", 00:15:27.695 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:27.695 "strip_size_kb": 0, 00:15:27.695 "state": "online", 00:15:27.695 "raid_level": "raid1", 00:15:27.695 "superblock": true, 00:15:27.695 "num_base_bdevs": 2, 00:15:27.695 "num_base_bdevs_discovered": 2, 00:15:27.695 "num_base_bdevs_operational": 2, 00:15:27.695 "base_bdevs_list": [ 00:15:27.695 { 00:15:27.695 "name": "spare", 00:15:27.695 "uuid": "abb70d3f-1d43-52d1-bee5-e651b9148266", 00:15:27.695 "is_configured": true, 00:15:27.695 "data_offset": 256, 00:15:27.695 "data_size": 7936 00:15:27.695 }, 00:15:27.695 { 00:15:27.695 "name": "BaseBdev2", 00:15:27.695 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:27.695 "is_configured": true, 00:15:27.695 "data_offset": 256, 00:15:27.695 "data_size": 7936 00:15:27.695 } 00:15:27.695 ] 00:15:27.695 }' 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:27.695 05:43:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.695 [2024-12-07 05:43:01.047565] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:27.695 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.955 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.955 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.956 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.956 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.956 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:27.956 "name": "raid_bdev1", 00:15:27.956 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:27.956 "strip_size_kb": 0, 00:15:27.956 "state": "online", 00:15:27.956 "raid_level": "raid1", 00:15:27.956 "superblock": true, 00:15:27.956 "num_base_bdevs": 2, 00:15:27.956 "num_base_bdevs_discovered": 1, 00:15:27.956 "num_base_bdevs_operational": 1, 00:15:27.956 "base_bdevs_list": [ 00:15:27.956 { 00:15:27.956 "name": null, 00:15:27.956 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:27.956 "is_configured": false, 00:15:27.956 "data_offset": 0, 00:15:27.956 "data_size": 7936 00:15:27.956 }, 00:15:27.956 { 00:15:27.956 "name": "BaseBdev2", 00:15:27.956 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:27.956 "is_configured": true, 00:15:27.956 "data_offset": 256, 00:15:27.956 "data_size": 7936 00:15:27.956 } 00:15:27.956 ] 00:15:27.956 }' 00:15:27.956 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:27.956 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:28.215 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:28.215 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.216 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:28.216 [2024-12-07 05:43:01.438895] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:28.216 [2024-12-07 05:43:01.439051] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:28.216 [2024-12-07 05:43:01.439066] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:28.216 [2024-12-07 05:43:01.439119] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:28.216 [2024-12-07 05:43:01.443860] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:15:28.216 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.216 05:43:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:28.216 [2024-12-07 05:43:01.445691] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:29.155 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:29.155 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:29.155 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:29.155 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:29.155 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:29.155 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:29.155 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.155 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.155 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:29.155 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.155 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:29.155 "name": "raid_bdev1", 00:15:29.155 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:29.155 "strip_size_kb": 0, 00:15:29.155 "state": "online", 00:15:29.155 "raid_level": "raid1", 00:15:29.155 "superblock": true, 00:15:29.155 "num_base_bdevs": 2, 00:15:29.155 "num_base_bdevs_discovered": 2, 00:15:29.155 "num_base_bdevs_operational": 2, 00:15:29.155 "process": { 00:15:29.155 "type": "rebuild", 00:15:29.155 "target": "spare", 00:15:29.155 "progress": { 00:15:29.155 "blocks": 2560, 00:15:29.155 "percent": 32 00:15:29.155 } 00:15:29.155 }, 00:15:29.155 "base_bdevs_list": [ 00:15:29.156 { 00:15:29.156 "name": "spare", 00:15:29.156 "uuid": "abb70d3f-1d43-52d1-bee5-e651b9148266", 00:15:29.156 "is_configured": true, 00:15:29.156 "data_offset": 256, 00:15:29.156 "data_size": 7936 00:15:29.156 }, 00:15:29.156 { 00:15:29.156 "name": "BaseBdev2", 00:15:29.156 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:29.156 "is_configured": true, 00:15:29.156 "data_offset": 256, 00:15:29.156 "data_size": 7936 00:15:29.156 } 00:15:29.156 ] 00:15:29.156 }' 00:15:29.156 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.416 [2024-12-07 05:43:02.582396] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:29.416 [2024-12-07 05:43:02.649558] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:29.416 [2024-12-07 05:43:02.649694] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:29.416 [2024-12-07 05:43:02.649735] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:29.416 [2024-12-07 05:43:02.649762] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:29.416 "name": "raid_bdev1", 00:15:29.416 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:29.416 "strip_size_kb": 0, 00:15:29.416 "state": "online", 00:15:29.416 "raid_level": "raid1", 00:15:29.416 "superblock": true, 00:15:29.416 "num_base_bdevs": 2, 00:15:29.416 "num_base_bdevs_discovered": 1, 00:15:29.416 "num_base_bdevs_operational": 1, 00:15:29.416 "base_bdevs_list": [ 00:15:29.416 { 00:15:29.416 "name": null, 00:15:29.416 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:29.416 "is_configured": false, 00:15:29.416 "data_offset": 0, 00:15:29.416 "data_size": 7936 00:15:29.416 }, 00:15:29.416 { 00:15:29.416 "name": "BaseBdev2", 00:15:29.416 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:29.416 "is_configured": true, 00:15:29.416 "data_offset": 256, 00:15:29.416 "data_size": 7936 00:15:29.416 } 00:15:29.416 ] 00:15:29.416 }' 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:29.416 05:43:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.987 05:43:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:29.987 05:43:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.987 05:43:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.987 [2024-12-07 05:43:03.085484] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:29.987 [2024-12-07 05:43:03.085541] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:29.987 [2024-12-07 05:43:03.085562] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:15:29.987 [2024-12-07 05:43:03.085572] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:29.987 [2024-12-07 05:43:03.086016] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:29.987 [2024-12-07 05:43:03.086041] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:29.987 [2024-12-07 05:43:03.086118] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:29.987 [2024-12-07 05:43:03.086130] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:29.987 [2024-12-07 05:43:03.086145] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:29.987 [2024-12-07 05:43:03.086172] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:29.987 [2024-12-07 05:43:03.090365] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:15:29.987 spare 00:15:29.987 05:43:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.987 05:43:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:29.987 [2024-12-07 05:43:03.092259] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:30.928 "name": "raid_bdev1", 00:15:30.928 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:30.928 "strip_size_kb": 0, 00:15:30.928 "state": "online", 00:15:30.928 "raid_level": "raid1", 00:15:30.928 "superblock": true, 00:15:30.928 "num_base_bdevs": 2, 00:15:30.928 "num_base_bdevs_discovered": 2, 00:15:30.928 "num_base_bdevs_operational": 2, 00:15:30.928 "process": { 00:15:30.928 "type": "rebuild", 00:15:30.928 "target": "spare", 00:15:30.928 "progress": { 00:15:30.928 "blocks": 2560, 00:15:30.928 "percent": 32 00:15:30.928 } 00:15:30.928 }, 00:15:30.928 "base_bdevs_list": [ 00:15:30.928 { 00:15:30.928 "name": "spare", 00:15:30.928 "uuid": "abb70d3f-1d43-52d1-bee5-e651b9148266", 00:15:30.928 "is_configured": true, 00:15:30.928 "data_offset": 256, 00:15:30.928 "data_size": 7936 00:15:30.928 }, 00:15:30.928 { 00:15:30.928 "name": "BaseBdev2", 00:15:30.928 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:30.928 "is_configured": true, 00:15:30.928 "data_offset": 256, 00:15:30.928 "data_size": 7936 00:15:30.928 } 00:15:30.928 ] 00:15:30.928 }' 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:30.928 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:30.928 [2024-12-07 05:43:04.236386] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:31.189 [2024-12-07 05:43:04.296208] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:31.189 [2024-12-07 05:43:04.296333] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:31.189 [2024-12-07 05:43:04.296369] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:31.189 [2024-12-07 05:43:04.296392] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:31.189 "name": "raid_bdev1", 00:15:31.189 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:31.189 "strip_size_kb": 0, 00:15:31.189 "state": "online", 00:15:31.189 "raid_level": "raid1", 00:15:31.189 "superblock": true, 00:15:31.189 "num_base_bdevs": 2, 00:15:31.189 "num_base_bdevs_discovered": 1, 00:15:31.189 "num_base_bdevs_operational": 1, 00:15:31.189 "base_bdevs_list": [ 00:15:31.189 { 00:15:31.189 "name": null, 00:15:31.189 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:31.189 "is_configured": false, 00:15:31.189 "data_offset": 0, 00:15:31.189 "data_size": 7936 00:15:31.189 }, 00:15:31.189 { 00:15:31.189 "name": "BaseBdev2", 00:15:31.189 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:31.189 "is_configured": true, 00:15:31.189 "data_offset": 256, 00:15:31.189 "data_size": 7936 00:15:31.189 } 00:15:31.189 ] 00:15:31.189 }' 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:31.189 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:31.452 "name": "raid_bdev1", 00:15:31.452 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:31.452 "strip_size_kb": 0, 00:15:31.452 "state": "online", 00:15:31.452 "raid_level": "raid1", 00:15:31.452 "superblock": true, 00:15:31.452 "num_base_bdevs": 2, 00:15:31.452 "num_base_bdevs_discovered": 1, 00:15:31.452 "num_base_bdevs_operational": 1, 00:15:31.452 "base_bdevs_list": [ 00:15:31.452 { 00:15:31.452 "name": null, 00:15:31.452 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:31.452 "is_configured": false, 00:15:31.452 "data_offset": 0, 00:15:31.452 "data_size": 7936 00:15:31.452 }, 00:15:31.452 { 00:15:31.452 "name": "BaseBdev2", 00:15:31.452 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:31.452 "is_configured": true, 00:15:31.452 "data_offset": 256, 00:15:31.452 "data_size": 7936 00:15:31.452 } 00:15:31.452 ] 00:15:31.452 }' 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:31.452 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:31.713 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:31.713 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:31.713 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.713 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.713 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.713 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:31.713 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.713 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.713 [2024-12-07 05:43:04.863768] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:31.713 [2024-12-07 05:43:04.863823] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:31.713 [2024-12-07 05:43:04.863841] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:15:31.713 [2024-12-07 05:43:04.863852] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:31.713 [2024-12-07 05:43:04.864236] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:31.713 [2024-12-07 05:43:04.864264] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:31.713 [2024-12-07 05:43:04.864357] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:31.713 [2024-12-07 05:43:04.864375] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:31.713 [2024-12-07 05:43:04.864388] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:31.713 [2024-12-07 05:43:04.864400] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:31.713 BaseBdev1 00:15:31.713 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.713 05:43:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:32.649 "name": "raid_bdev1", 00:15:32.649 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:32.649 "strip_size_kb": 0, 00:15:32.649 "state": "online", 00:15:32.649 "raid_level": "raid1", 00:15:32.649 "superblock": true, 00:15:32.649 "num_base_bdevs": 2, 00:15:32.649 "num_base_bdevs_discovered": 1, 00:15:32.649 "num_base_bdevs_operational": 1, 00:15:32.649 "base_bdevs_list": [ 00:15:32.649 { 00:15:32.649 "name": null, 00:15:32.649 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:32.649 "is_configured": false, 00:15:32.649 "data_offset": 0, 00:15:32.649 "data_size": 7936 00:15:32.649 }, 00:15:32.649 { 00:15:32.649 "name": "BaseBdev2", 00:15:32.649 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:32.649 "is_configured": true, 00:15:32.649 "data_offset": 256, 00:15:32.649 "data_size": 7936 00:15:32.649 } 00:15:32.649 ] 00:15:32.649 }' 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:32.649 05:43:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.908 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:32.908 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:32.908 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:32.908 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:32.908 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:32.908 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.908 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.908 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.908 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:33.167 "name": "raid_bdev1", 00:15:33.167 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:33.167 "strip_size_kb": 0, 00:15:33.167 "state": "online", 00:15:33.167 "raid_level": "raid1", 00:15:33.167 "superblock": true, 00:15:33.167 "num_base_bdevs": 2, 00:15:33.167 "num_base_bdevs_discovered": 1, 00:15:33.167 "num_base_bdevs_operational": 1, 00:15:33.167 "base_bdevs_list": [ 00:15:33.167 { 00:15:33.167 "name": null, 00:15:33.167 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:33.167 "is_configured": false, 00:15:33.167 "data_offset": 0, 00:15:33.167 "data_size": 7936 00:15:33.167 }, 00:15:33.167 { 00:15:33.167 "name": "BaseBdev2", 00:15:33.167 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:33.167 "is_configured": true, 00:15:33.167 "data_offset": 256, 00:15:33.167 "data_size": 7936 00:15:33.167 } 00:15:33.167 ] 00:15:33.167 }' 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # local es=0 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.167 [2024-12-07 05:43:06.393083] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:33.167 [2024-12-07 05:43:06.393308] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:33.167 [2024-12-07 05:43:06.393368] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:33.167 request: 00:15:33.167 { 00:15:33.167 "base_bdev": "BaseBdev1", 00:15:33.167 "raid_bdev": "raid_bdev1", 00:15:33.167 "method": "bdev_raid_add_base_bdev", 00:15:33.167 "req_id": 1 00:15:33.167 } 00:15:33.167 Got JSON-RPC error response 00:15:33.167 response: 00:15:33.167 { 00:15:33.167 "code": -22, 00:15:33.167 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:33.167 } 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # es=1 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:33.167 05:43:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:34.107 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:34.107 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:34.107 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:34.107 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:34.107 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:34.107 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:34.107 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:34.107 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:34.108 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:34.108 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:34.108 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.108 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.108 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.108 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.108 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.108 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:34.108 "name": "raid_bdev1", 00:15:34.108 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:34.108 "strip_size_kb": 0, 00:15:34.108 "state": "online", 00:15:34.108 "raid_level": "raid1", 00:15:34.108 "superblock": true, 00:15:34.108 "num_base_bdevs": 2, 00:15:34.108 "num_base_bdevs_discovered": 1, 00:15:34.108 "num_base_bdevs_operational": 1, 00:15:34.108 "base_bdevs_list": [ 00:15:34.108 { 00:15:34.108 "name": null, 00:15:34.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:34.108 "is_configured": false, 00:15:34.108 "data_offset": 0, 00:15:34.108 "data_size": 7936 00:15:34.108 }, 00:15:34.108 { 00:15:34.108 "name": "BaseBdev2", 00:15:34.108 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:34.108 "is_configured": true, 00:15:34.108 "data_offset": 256, 00:15:34.108 "data_size": 7936 00:15:34.108 } 00:15:34.108 ] 00:15:34.108 }' 00:15:34.108 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:34.108 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:34.678 "name": "raid_bdev1", 00:15:34.678 "uuid": "3deb3884-1b7a-4ef5-ba09-30954c7c904b", 00:15:34.678 "strip_size_kb": 0, 00:15:34.678 "state": "online", 00:15:34.678 "raid_level": "raid1", 00:15:34.678 "superblock": true, 00:15:34.678 "num_base_bdevs": 2, 00:15:34.678 "num_base_bdevs_discovered": 1, 00:15:34.678 "num_base_bdevs_operational": 1, 00:15:34.678 "base_bdevs_list": [ 00:15:34.678 { 00:15:34.678 "name": null, 00:15:34.678 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:34.678 "is_configured": false, 00:15:34.678 "data_offset": 0, 00:15:34.678 "data_size": 7936 00:15:34.678 }, 00:15:34.678 { 00:15:34.678 "name": "BaseBdev2", 00:15:34.678 "uuid": "52424fe9-7c15-5d6f-ba6e-e3060aab454d", 00:15:34.678 "is_configured": true, 00:15:34.678 "data_offset": 256, 00:15:34.678 "data_size": 7936 00:15:34.678 } 00:15:34.678 ] 00:15:34.678 }' 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 96446 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 96446 ']' 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 96446 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96446 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:34.678 killing process with pid 96446 00:15:34.678 Received shutdown signal, test time was about 60.000000 seconds 00:15:34.678 00:15:34.678 Latency(us) 00:15:34.678 [2024-12-07T05:43:08.046Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:34.678 [2024-12-07T05:43:08.046Z] =================================================================================================================== 00:15:34.678 [2024-12-07T05:43:08.046Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96446' 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@973 -- # kill 96446 00:15:34.678 [2024-12-07 05:43:07.977531] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:34.678 [2024-12-07 05:43:07.977656] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:34.678 [2024-12-07 05:43:07.977709] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:34.678 [2024-12-07 05:43:07.977719] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:15:34.678 05:43:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@978 -- # wait 96446 00:15:34.678 [2024-12-07 05:43:08.008175] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:34.939 05:43:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:15:34.939 00:15:34.939 real 0m18.045s 00:15:34.939 user 0m23.818s 00:15:34.939 sys 0m2.484s 00:15:34.939 ************************************ 00:15:34.939 END TEST raid_rebuild_test_sb_4k 00:15:34.939 ************************************ 00:15:34.939 05:43:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:34.939 05:43:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.939 05:43:08 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:15:34.939 05:43:08 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:15:34.939 05:43:08 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:15:34.939 05:43:08 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:34.939 05:43:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:34.939 ************************************ 00:15:34.939 START TEST raid_state_function_test_sb_md_separate 00:15:34.939 ************************************ 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=97121 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 97121' 00:15:34.939 Process raid pid: 97121 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 97121 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 97121 ']' 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:34.939 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:34.939 05:43:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:35.199 [2024-12-07 05:43:08.370123] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:15:35.199 [2024-12-07 05:43:08.370242] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:35.199 [2024-12-07 05:43:08.502188] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:35.199 [2024-12-07 05:43:08.526744] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:35.459 [2024-12-07 05:43:08.568924] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:35.459 [2024-12-07 05:43:08.568961] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.029 [2024-12-07 05:43:09.195432] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:36.029 [2024-12-07 05:43:09.195591] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:36.029 [2024-12-07 05:43:09.195635] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:36.029 [2024-12-07 05:43:09.195662] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:36.029 "name": "Existed_Raid", 00:15:36.029 "uuid": "09569d3d-5cb1-4570-baf0-2c02adad7704", 00:15:36.029 "strip_size_kb": 0, 00:15:36.029 "state": "configuring", 00:15:36.029 "raid_level": "raid1", 00:15:36.029 "superblock": true, 00:15:36.029 "num_base_bdevs": 2, 00:15:36.029 "num_base_bdevs_discovered": 0, 00:15:36.029 "num_base_bdevs_operational": 2, 00:15:36.029 "base_bdevs_list": [ 00:15:36.029 { 00:15:36.029 "name": "BaseBdev1", 00:15:36.029 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:36.029 "is_configured": false, 00:15:36.029 "data_offset": 0, 00:15:36.029 "data_size": 0 00:15:36.029 }, 00:15:36.029 { 00:15:36.029 "name": "BaseBdev2", 00:15:36.029 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:36.029 "is_configured": false, 00:15:36.029 "data_offset": 0, 00:15:36.029 "data_size": 0 00:15:36.029 } 00:15:36.029 ] 00:15:36.029 }' 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:36.029 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.289 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:36.289 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.289 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.289 [2024-12-07 05:43:09.646569] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:36.289 [2024-12-07 05:43:09.646685] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:15:36.289 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.289 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.550 [2024-12-07 05:43:09.658560] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:36.550 [2024-12-07 05:43:09.658675] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:36.550 [2024-12-07 05:43:09.658708] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:36.550 [2024-12-07 05:43:09.658743] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.550 [2024-12-07 05:43:09.679906] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:36.550 BaseBdev1 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.550 [ 00:15:36.550 { 00:15:36.550 "name": "BaseBdev1", 00:15:36.550 "aliases": [ 00:15:36.550 "0a14c7b5-269e-4f18-a177-b065f8707881" 00:15:36.550 ], 00:15:36.550 "product_name": "Malloc disk", 00:15:36.550 "block_size": 4096, 00:15:36.550 "num_blocks": 8192, 00:15:36.550 "uuid": "0a14c7b5-269e-4f18-a177-b065f8707881", 00:15:36.550 "md_size": 32, 00:15:36.550 "md_interleave": false, 00:15:36.550 "dif_type": 0, 00:15:36.550 "assigned_rate_limits": { 00:15:36.550 "rw_ios_per_sec": 0, 00:15:36.550 "rw_mbytes_per_sec": 0, 00:15:36.550 "r_mbytes_per_sec": 0, 00:15:36.550 "w_mbytes_per_sec": 0 00:15:36.550 }, 00:15:36.550 "claimed": true, 00:15:36.550 "claim_type": "exclusive_write", 00:15:36.550 "zoned": false, 00:15:36.550 "supported_io_types": { 00:15:36.550 "read": true, 00:15:36.550 "write": true, 00:15:36.550 "unmap": true, 00:15:36.550 "flush": true, 00:15:36.550 "reset": true, 00:15:36.550 "nvme_admin": false, 00:15:36.550 "nvme_io": false, 00:15:36.550 "nvme_io_md": false, 00:15:36.550 "write_zeroes": true, 00:15:36.550 "zcopy": true, 00:15:36.550 "get_zone_info": false, 00:15:36.550 "zone_management": false, 00:15:36.550 "zone_append": false, 00:15:36.550 "compare": false, 00:15:36.550 "compare_and_write": false, 00:15:36.550 "abort": true, 00:15:36.550 "seek_hole": false, 00:15:36.550 "seek_data": false, 00:15:36.550 "copy": true, 00:15:36.550 "nvme_iov_md": false 00:15:36.550 }, 00:15:36.550 "memory_domains": [ 00:15:36.550 { 00:15:36.550 "dma_device_id": "system", 00:15:36.550 "dma_device_type": 1 00:15:36.550 }, 00:15:36.550 { 00:15:36.550 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:36.550 "dma_device_type": 2 00:15:36.550 } 00:15:36.550 ], 00:15:36.550 "driver_specific": {} 00:15:36.550 } 00:15:36.550 ] 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:36.550 "name": "Existed_Raid", 00:15:36.550 "uuid": "5569d450-5af9-4574-babb-fdd564b58ab8", 00:15:36.550 "strip_size_kb": 0, 00:15:36.550 "state": "configuring", 00:15:36.550 "raid_level": "raid1", 00:15:36.550 "superblock": true, 00:15:36.550 "num_base_bdevs": 2, 00:15:36.550 "num_base_bdevs_discovered": 1, 00:15:36.550 "num_base_bdevs_operational": 2, 00:15:36.550 "base_bdevs_list": [ 00:15:36.550 { 00:15:36.550 "name": "BaseBdev1", 00:15:36.550 "uuid": "0a14c7b5-269e-4f18-a177-b065f8707881", 00:15:36.550 "is_configured": true, 00:15:36.550 "data_offset": 256, 00:15:36.550 "data_size": 7936 00:15:36.550 }, 00:15:36.550 { 00:15:36.550 "name": "BaseBdev2", 00:15:36.550 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:36.550 "is_configured": false, 00:15:36.550 "data_offset": 0, 00:15:36.550 "data_size": 0 00:15:36.550 } 00:15:36.550 ] 00:15:36.550 }' 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:36.550 05:43:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.810 [2024-12-07 05:43:10.143179] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:36.810 [2024-12-07 05:43:10.143269] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:36.810 [2024-12-07 05:43:10.155206] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:36.810 [2024-12-07 05:43:10.157056] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:36.810 [2024-12-07 05:43:10.157145] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:36.810 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:37.070 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.070 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:37.070 "name": "Existed_Raid", 00:15:37.070 "uuid": "8349c0eb-e0b2-411e-a0df-d24099fe7048", 00:15:37.070 "strip_size_kb": 0, 00:15:37.070 "state": "configuring", 00:15:37.070 "raid_level": "raid1", 00:15:37.070 "superblock": true, 00:15:37.070 "num_base_bdevs": 2, 00:15:37.070 "num_base_bdevs_discovered": 1, 00:15:37.070 "num_base_bdevs_operational": 2, 00:15:37.070 "base_bdevs_list": [ 00:15:37.070 { 00:15:37.070 "name": "BaseBdev1", 00:15:37.070 "uuid": "0a14c7b5-269e-4f18-a177-b065f8707881", 00:15:37.070 "is_configured": true, 00:15:37.070 "data_offset": 256, 00:15:37.070 "data_size": 7936 00:15:37.070 }, 00:15:37.070 { 00:15:37.070 "name": "BaseBdev2", 00:15:37.070 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:37.070 "is_configured": false, 00:15:37.070 "data_offset": 0, 00:15:37.070 "data_size": 0 00:15:37.070 } 00:15:37.070 ] 00:15:37.070 }' 00:15:37.070 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:37.070 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:37.330 [2024-12-07 05:43:10.597965] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:37.330 [2024-12-07 05:43:10.598252] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:37.330 [2024-12-07 05:43:10.598313] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:37.330 [2024-12-07 05:43:10.598442] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:37.330 [2024-12-07 05:43:10.598600] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:37.330 [2024-12-07 05:43:10.598673] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:15:37.330 [2024-12-07 05:43:10.598795] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:37.330 BaseBdev2 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:37.330 [ 00:15:37.330 { 00:15:37.330 "name": "BaseBdev2", 00:15:37.330 "aliases": [ 00:15:37.330 "5290573b-b6ce-4be4-b4dd-643e2098e442" 00:15:37.330 ], 00:15:37.330 "product_name": "Malloc disk", 00:15:37.330 "block_size": 4096, 00:15:37.330 "num_blocks": 8192, 00:15:37.330 "uuid": "5290573b-b6ce-4be4-b4dd-643e2098e442", 00:15:37.330 "md_size": 32, 00:15:37.330 "md_interleave": false, 00:15:37.330 "dif_type": 0, 00:15:37.330 "assigned_rate_limits": { 00:15:37.330 "rw_ios_per_sec": 0, 00:15:37.330 "rw_mbytes_per_sec": 0, 00:15:37.330 "r_mbytes_per_sec": 0, 00:15:37.330 "w_mbytes_per_sec": 0 00:15:37.330 }, 00:15:37.330 "claimed": true, 00:15:37.330 "claim_type": "exclusive_write", 00:15:37.330 "zoned": false, 00:15:37.330 "supported_io_types": { 00:15:37.330 "read": true, 00:15:37.330 "write": true, 00:15:37.330 "unmap": true, 00:15:37.330 "flush": true, 00:15:37.330 "reset": true, 00:15:37.330 "nvme_admin": false, 00:15:37.330 "nvme_io": false, 00:15:37.330 "nvme_io_md": false, 00:15:37.330 "write_zeroes": true, 00:15:37.330 "zcopy": true, 00:15:37.330 "get_zone_info": false, 00:15:37.330 "zone_management": false, 00:15:37.330 "zone_append": false, 00:15:37.330 "compare": false, 00:15:37.330 "compare_and_write": false, 00:15:37.330 "abort": true, 00:15:37.330 "seek_hole": false, 00:15:37.330 "seek_data": false, 00:15:37.330 "copy": true, 00:15:37.330 "nvme_iov_md": false 00:15:37.330 }, 00:15:37.330 "memory_domains": [ 00:15:37.330 { 00:15:37.330 "dma_device_id": "system", 00:15:37.330 "dma_device_type": 1 00:15:37.330 }, 00:15:37.330 { 00:15:37.330 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:37.330 "dma_device_type": 2 00:15:37.330 } 00:15:37.330 ], 00:15:37.330 "driver_specific": {} 00:15:37.330 } 00:15:37.330 ] 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:37.330 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:37.331 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.591 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:37.591 "name": "Existed_Raid", 00:15:37.591 "uuid": "8349c0eb-e0b2-411e-a0df-d24099fe7048", 00:15:37.591 "strip_size_kb": 0, 00:15:37.591 "state": "online", 00:15:37.591 "raid_level": "raid1", 00:15:37.591 "superblock": true, 00:15:37.591 "num_base_bdevs": 2, 00:15:37.591 "num_base_bdevs_discovered": 2, 00:15:37.591 "num_base_bdevs_operational": 2, 00:15:37.591 "base_bdevs_list": [ 00:15:37.591 { 00:15:37.591 "name": "BaseBdev1", 00:15:37.591 "uuid": "0a14c7b5-269e-4f18-a177-b065f8707881", 00:15:37.591 "is_configured": true, 00:15:37.591 "data_offset": 256, 00:15:37.591 "data_size": 7936 00:15:37.591 }, 00:15:37.591 { 00:15:37.591 "name": "BaseBdev2", 00:15:37.591 "uuid": "5290573b-b6ce-4be4-b4dd-643e2098e442", 00:15:37.591 "is_configured": true, 00:15:37.591 "data_offset": 256, 00:15:37.591 "data_size": 7936 00:15:37.591 } 00:15:37.591 ] 00:15:37.591 }' 00:15:37.591 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:37.591 05:43:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:37.851 [2024-12-07 05:43:11.069465] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.851 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:37.851 "name": "Existed_Raid", 00:15:37.851 "aliases": [ 00:15:37.851 "8349c0eb-e0b2-411e-a0df-d24099fe7048" 00:15:37.851 ], 00:15:37.851 "product_name": "Raid Volume", 00:15:37.851 "block_size": 4096, 00:15:37.851 "num_blocks": 7936, 00:15:37.851 "uuid": "8349c0eb-e0b2-411e-a0df-d24099fe7048", 00:15:37.851 "md_size": 32, 00:15:37.851 "md_interleave": false, 00:15:37.851 "dif_type": 0, 00:15:37.851 "assigned_rate_limits": { 00:15:37.851 "rw_ios_per_sec": 0, 00:15:37.851 "rw_mbytes_per_sec": 0, 00:15:37.851 "r_mbytes_per_sec": 0, 00:15:37.851 "w_mbytes_per_sec": 0 00:15:37.851 }, 00:15:37.851 "claimed": false, 00:15:37.851 "zoned": false, 00:15:37.851 "supported_io_types": { 00:15:37.851 "read": true, 00:15:37.851 "write": true, 00:15:37.851 "unmap": false, 00:15:37.851 "flush": false, 00:15:37.851 "reset": true, 00:15:37.851 "nvme_admin": false, 00:15:37.851 "nvme_io": false, 00:15:37.851 "nvme_io_md": false, 00:15:37.851 "write_zeroes": true, 00:15:37.851 "zcopy": false, 00:15:37.851 "get_zone_info": false, 00:15:37.851 "zone_management": false, 00:15:37.851 "zone_append": false, 00:15:37.851 "compare": false, 00:15:37.851 "compare_and_write": false, 00:15:37.851 "abort": false, 00:15:37.851 "seek_hole": false, 00:15:37.851 "seek_data": false, 00:15:37.851 "copy": false, 00:15:37.851 "nvme_iov_md": false 00:15:37.851 }, 00:15:37.851 "memory_domains": [ 00:15:37.851 { 00:15:37.851 "dma_device_id": "system", 00:15:37.851 "dma_device_type": 1 00:15:37.851 }, 00:15:37.851 { 00:15:37.851 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:37.851 "dma_device_type": 2 00:15:37.851 }, 00:15:37.851 { 00:15:37.851 "dma_device_id": "system", 00:15:37.851 "dma_device_type": 1 00:15:37.852 }, 00:15:37.852 { 00:15:37.852 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:37.852 "dma_device_type": 2 00:15:37.852 } 00:15:37.852 ], 00:15:37.852 "driver_specific": { 00:15:37.852 "raid": { 00:15:37.852 "uuid": "8349c0eb-e0b2-411e-a0df-d24099fe7048", 00:15:37.852 "strip_size_kb": 0, 00:15:37.852 "state": "online", 00:15:37.852 "raid_level": "raid1", 00:15:37.852 "superblock": true, 00:15:37.852 "num_base_bdevs": 2, 00:15:37.852 "num_base_bdevs_discovered": 2, 00:15:37.852 "num_base_bdevs_operational": 2, 00:15:37.852 "base_bdevs_list": [ 00:15:37.852 { 00:15:37.852 "name": "BaseBdev1", 00:15:37.852 "uuid": "0a14c7b5-269e-4f18-a177-b065f8707881", 00:15:37.852 "is_configured": true, 00:15:37.852 "data_offset": 256, 00:15:37.852 "data_size": 7936 00:15:37.852 }, 00:15:37.852 { 00:15:37.852 "name": "BaseBdev2", 00:15:37.852 "uuid": "5290573b-b6ce-4be4-b4dd-643e2098e442", 00:15:37.852 "is_configured": true, 00:15:37.852 "data_offset": 256, 00:15:37.852 "data_size": 7936 00:15:37.852 } 00:15:37.852 ] 00:15:37.852 } 00:15:37.852 } 00:15:37.852 }' 00:15:37.852 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:37.852 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:37.852 BaseBdev2' 00:15:37.852 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:37.852 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:37.852 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:37.852 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:37.852 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:37.852 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.852 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:37.852 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:38.112 [2024-12-07 05:43:11.276920] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:38.112 "name": "Existed_Raid", 00:15:38.112 "uuid": "8349c0eb-e0b2-411e-a0df-d24099fe7048", 00:15:38.112 "strip_size_kb": 0, 00:15:38.112 "state": "online", 00:15:38.112 "raid_level": "raid1", 00:15:38.112 "superblock": true, 00:15:38.112 "num_base_bdevs": 2, 00:15:38.112 "num_base_bdevs_discovered": 1, 00:15:38.112 "num_base_bdevs_operational": 1, 00:15:38.112 "base_bdevs_list": [ 00:15:38.112 { 00:15:38.112 "name": null, 00:15:38.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.112 "is_configured": false, 00:15:38.112 "data_offset": 0, 00:15:38.112 "data_size": 7936 00:15:38.112 }, 00:15:38.112 { 00:15:38.112 "name": "BaseBdev2", 00:15:38.112 "uuid": "5290573b-b6ce-4be4-b4dd-643e2098e442", 00:15:38.112 "is_configured": true, 00:15:38.112 "data_offset": 256, 00:15:38.112 "data_size": 7936 00:15:38.112 } 00:15:38.112 ] 00:15:38.112 }' 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:38.112 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.372 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:38.372 [2024-12-07 05:43:11.736118] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:38.372 [2024-12-07 05:43:11.736286] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:38.632 [2024-12-07 05:43:11.748214] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:38.632 [2024-12-07 05:43:11.748322] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:38.632 [2024-12-07 05:43:11.748366] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 97121 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 97121 ']' 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 97121 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97121 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:38.632 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:38.633 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97121' 00:15:38.633 killing process with pid 97121 00:15:38.633 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 97121 00:15:38.633 [2024-12-07 05:43:11.841158] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:38.633 05:43:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 97121 00:15:38.633 [2024-12-07 05:43:11.842113] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:38.892 05:43:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:15:38.892 ************************************ 00:15:38.892 END TEST raid_state_function_test_sb_md_separate 00:15:38.892 ************************************ 00:15:38.892 00:15:38.892 real 0m3.768s 00:15:38.892 user 0m5.898s 00:15:38.892 sys 0m0.806s 00:15:38.892 05:43:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:38.892 05:43:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:38.892 05:43:12 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:15:38.892 05:43:12 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:15:38.892 05:43:12 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:38.892 05:43:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:38.892 ************************************ 00:15:38.892 START TEST raid_superblock_test_md_separate 00:15:38.892 ************************************ 00:15:38.892 05:43:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:15:38.892 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:15:38.892 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:15:38.892 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:38.892 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:38.892 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:38.892 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=97357 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 97357 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@835 -- # '[' -z 97357 ']' 00:15:38.893 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:38.893 05:43:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:38.893 [2024-12-07 05:43:12.204765] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:15:38.893 [2024-12-07 05:43:12.204879] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97357 ] 00:15:39.152 [2024-12-07 05:43:12.336460] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:39.152 [2024-12-07 05:43:12.360879] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:39.152 [2024-12-07 05:43:12.402804] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:39.152 [2024-12-07 05:43:12.402843] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@868 -- # return 0 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:15:39.722 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:39.723 malloc1 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:39.723 [2024-12-07 05:43:13.042277] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:39.723 [2024-12-07 05:43:13.042353] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:39.723 [2024-12-07 05:43:13.042375] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:39.723 [2024-12-07 05:43:13.042392] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:39.723 [2024-12-07 05:43:13.044287] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:39.723 [2024-12-07 05:43:13.044372] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:39.723 pt1 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:39.723 malloc2 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:39.723 [2024-12-07 05:43:13.071287] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:39.723 [2024-12-07 05:43:13.071415] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:39.723 [2024-12-07 05:43:13.071449] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:39.723 [2024-12-07 05:43:13.071481] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:39.723 [2024-12-07 05:43:13.073334] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:39.723 [2024-12-07 05:43:13.073430] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:39.723 pt2 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:39.723 [2024-12-07 05:43:13.083302] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:39.723 [2024-12-07 05:43:13.085138] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:39.723 [2024-12-07 05:43:13.085339] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:39.723 [2024-12-07 05:43:13.085393] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:39.723 [2024-12-07 05:43:13.085501] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:39.723 [2024-12-07 05:43:13.085664] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:39.723 [2024-12-07 05:43:13.085708] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:39.723 [2024-12-07 05:43:13.085833] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.723 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:39.983 "name": "raid_bdev1", 00:15:39.983 "uuid": "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369", 00:15:39.983 "strip_size_kb": 0, 00:15:39.983 "state": "online", 00:15:39.983 "raid_level": "raid1", 00:15:39.983 "superblock": true, 00:15:39.983 "num_base_bdevs": 2, 00:15:39.983 "num_base_bdevs_discovered": 2, 00:15:39.983 "num_base_bdevs_operational": 2, 00:15:39.983 "base_bdevs_list": [ 00:15:39.983 { 00:15:39.983 "name": "pt1", 00:15:39.983 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:39.983 "is_configured": true, 00:15:39.983 "data_offset": 256, 00:15:39.983 "data_size": 7936 00:15:39.983 }, 00:15:39.983 { 00:15:39.983 "name": "pt2", 00:15:39.983 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:39.983 "is_configured": true, 00:15:39.983 "data_offset": 256, 00:15:39.983 "data_size": 7936 00:15:39.983 } 00:15:39.983 ] 00:15:39.983 }' 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:39.983 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.252 [2024-12-07 05:43:13.518824] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.252 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:40.252 "name": "raid_bdev1", 00:15:40.252 "aliases": [ 00:15:40.252 "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369" 00:15:40.252 ], 00:15:40.252 "product_name": "Raid Volume", 00:15:40.252 "block_size": 4096, 00:15:40.252 "num_blocks": 7936, 00:15:40.252 "uuid": "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369", 00:15:40.252 "md_size": 32, 00:15:40.252 "md_interleave": false, 00:15:40.252 "dif_type": 0, 00:15:40.252 "assigned_rate_limits": { 00:15:40.252 "rw_ios_per_sec": 0, 00:15:40.252 "rw_mbytes_per_sec": 0, 00:15:40.252 "r_mbytes_per_sec": 0, 00:15:40.252 "w_mbytes_per_sec": 0 00:15:40.252 }, 00:15:40.252 "claimed": false, 00:15:40.252 "zoned": false, 00:15:40.252 "supported_io_types": { 00:15:40.252 "read": true, 00:15:40.252 "write": true, 00:15:40.252 "unmap": false, 00:15:40.252 "flush": false, 00:15:40.252 "reset": true, 00:15:40.252 "nvme_admin": false, 00:15:40.252 "nvme_io": false, 00:15:40.252 "nvme_io_md": false, 00:15:40.252 "write_zeroes": true, 00:15:40.252 "zcopy": false, 00:15:40.252 "get_zone_info": false, 00:15:40.252 "zone_management": false, 00:15:40.252 "zone_append": false, 00:15:40.252 "compare": false, 00:15:40.252 "compare_and_write": false, 00:15:40.252 "abort": false, 00:15:40.252 "seek_hole": false, 00:15:40.252 "seek_data": false, 00:15:40.252 "copy": false, 00:15:40.252 "nvme_iov_md": false 00:15:40.252 }, 00:15:40.252 "memory_domains": [ 00:15:40.252 { 00:15:40.252 "dma_device_id": "system", 00:15:40.252 "dma_device_type": 1 00:15:40.252 }, 00:15:40.252 { 00:15:40.252 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:40.252 "dma_device_type": 2 00:15:40.252 }, 00:15:40.252 { 00:15:40.252 "dma_device_id": "system", 00:15:40.252 "dma_device_type": 1 00:15:40.252 }, 00:15:40.252 { 00:15:40.252 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:40.252 "dma_device_type": 2 00:15:40.252 } 00:15:40.252 ], 00:15:40.252 "driver_specific": { 00:15:40.252 "raid": { 00:15:40.252 "uuid": "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369", 00:15:40.252 "strip_size_kb": 0, 00:15:40.252 "state": "online", 00:15:40.252 "raid_level": "raid1", 00:15:40.252 "superblock": true, 00:15:40.252 "num_base_bdevs": 2, 00:15:40.252 "num_base_bdevs_discovered": 2, 00:15:40.252 "num_base_bdevs_operational": 2, 00:15:40.252 "base_bdevs_list": [ 00:15:40.252 { 00:15:40.252 "name": "pt1", 00:15:40.252 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:40.252 "is_configured": true, 00:15:40.252 "data_offset": 256, 00:15:40.252 "data_size": 7936 00:15:40.252 }, 00:15:40.252 { 00:15:40.253 "name": "pt2", 00:15:40.253 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:40.253 "is_configured": true, 00:15:40.253 "data_offset": 256, 00:15:40.253 "data_size": 7936 00:15:40.253 } 00:15:40.253 ] 00:15:40.253 } 00:15:40.253 } 00:15:40.253 }' 00:15:40.253 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:40.253 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:40.253 pt2' 00:15:40.253 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.532 [2024-12-07 05:43:13.746361] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369 ']' 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.532 [2024-12-07 05:43:13.794074] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:40.532 [2024-12-07 05:43:13.794152] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:40.532 [2024-12-07 05:43:13.794254] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:40.532 [2024-12-07 05:43:13.794359] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:40.532 [2024-12-07 05:43:13.794433] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:40.532 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.810 [2024-12-07 05:43:13.933836] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:40.810 [2024-12-07 05:43:13.935693] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:40.810 [2024-12-07 05:43:13.935814] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:40.810 [2024-12-07 05:43:13.935892] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:40.810 [2024-12-07 05:43:13.935933] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:40.810 [2024-12-07 05:43:13.935964] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:15:40.810 request: 00:15:40.810 { 00:15:40.810 "name": "raid_bdev1", 00:15:40.810 "raid_level": "raid1", 00:15:40.810 "base_bdevs": [ 00:15:40.810 "malloc1", 00:15:40.810 "malloc2" 00:15:40.810 ], 00:15:40.810 "superblock": false, 00:15:40.810 "method": "bdev_raid_create", 00:15:40.810 "req_id": 1 00:15:40.810 } 00:15:40.810 Got JSON-RPC error response 00:15:40.810 response: 00:15:40.810 { 00:15:40.810 "code": -17, 00:15:40.810 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:40.810 } 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # es=1 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.810 [2024-12-07 05:43:13.985713] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:40.810 [2024-12-07 05:43:13.985821] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:40.810 [2024-12-07 05:43:13.985854] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:40.810 [2024-12-07 05:43:13.985881] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:40.810 [2024-12-07 05:43:13.987734] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:40.810 [2024-12-07 05:43:13.987804] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:40.810 [2024-12-07 05:43:13.987867] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:40.810 [2024-12-07 05:43:13.987936] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:40.810 pt1 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:40.810 05:43:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:40.810 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.810 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:40.810 "name": "raid_bdev1", 00:15:40.810 "uuid": "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369", 00:15:40.810 "strip_size_kb": 0, 00:15:40.810 "state": "configuring", 00:15:40.810 "raid_level": "raid1", 00:15:40.810 "superblock": true, 00:15:40.810 "num_base_bdevs": 2, 00:15:40.810 "num_base_bdevs_discovered": 1, 00:15:40.810 "num_base_bdevs_operational": 2, 00:15:40.810 "base_bdevs_list": [ 00:15:40.810 { 00:15:40.810 "name": "pt1", 00:15:40.810 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:40.810 "is_configured": true, 00:15:40.810 "data_offset": 256, 00:15:40.810 "data_size": 7936 00:15:40.810 }, 00:15:40.810 { 00:15:40.810 "name": null, 00:15:40.810 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:40.810 "is_configured": false, 00:15:40.811 "data_offset": 256, 00:15:40.811 "data_size": 7936 00:15:40.811 } 00:15:40.811 ] 00:15:40.811 }' 00:15:40.811 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:40.811 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:41.070 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:15:41.070 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:41.070 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:41.070 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:41.070 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.070 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:41.070 [2024-12-07 05:43:14.420950] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:41.070 [2024-12-07 05:43:14.421046] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:41.070 [2024-12-07 05:43:14.421081] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:41.070 [2024-12-07 05:43:14.421108] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:41.070 [2024-12-07 05:43:14.421284] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:41.070 [2024-12-07 05:43:14.421348] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:41.070 [2024-12-07 05:43:14.421417] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:41.070 [2024-12-07 05:43:14.421471] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:41.070 [2024-12-07 05:43:14.421587] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:41.070 [2024-12-07 05:43:14.421621] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:41.070 [2024-12-07 05:43:14.421718] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:41.070 [2024-12-07 05:43:14.421805] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:41.071 [2024-12-07 05:43:14.421818] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:15:41.071 [2024-12-07 05:43:14.421879] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:41.071 pt2 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.071 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:41.330 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.330 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:41.330 "name": "raid_bdev1", 00:15:41.330 "uuid": "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369", 00:15:41.330 "strip_size_kb": 0, 00:15:41.330 "state": "online", 00:15:41.330 "raid_level": "raid1", 00:15:41.330 "superblock": true, 00:15:41.330 "num_base_bdevs": 2, 00:15:41.330 "num_base_bdevs_discovered": 2, 00:15:41.330 "num_base_bdevs_operational": 2, 00:15:41.330 "base_bdevs_list": [ 00:15:41.330 { 00:15:41.330 "name": "pt1", 00:15:41.330 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:41.330 "is_configured": true, 00:15:41.330 "data_offset": 256, 00:15:41.330 "data_size": 7936 00:15:41.330 }, 00:15:41.330 { 00:15:41.330 "name": "pt2", 00:15:41.330 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:41.330 "is_configured": true, 00:15:41.330 "data_offset": 256, 00:15:41.330 "data_size": 7936 00:15:41.330 } 00:15:41.330 ] 00:15:41.330 }' 00:15:41.330 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:41.330 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:41.590 [2024-12-07 05:43:14.844507] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.590 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:41.590 "name": "raid_bdev1", 00:15:41.590 "aliases": [ 00:15:41.590 "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369" 00:15:41.590 ], 00:15:41.590 "product_name": "Raid Volume", 00:15:41.590 "block_size": 4096, 00:15:41.590 "num_blocks": 7936, 00:15:41.590 "uuid": "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369", 00:15:41.590 "md_size": 32, 00:15:41.590 "md_interleave": false, 00:15:41.590 "dif_type": 0, 00:15:41.590 "assigned_rate_limits": { 00:15:41.590 "rw_ios_per_sec": 0, 00:15:41.590 "rw_mbytes_per_sec": 0, 00:15:41.590 "r_mbytes_per_sec": 0, 00:15:41.590 "w_mbytes_per_sec": 0 00:15:41.590 }, 00:15:41.590 "claimed": false, 00:15:41.590 "zoned": false, 00:15:41.590 "supported_io_types": { 00:15:41.590 "read": true, 00:15:41.590 "write": true, 00:15:41.590 "unmap": false, 00:15:41.590 "flush": false, 00:15:41.590 "reset": true, 00:15:41.590 "nvme_admin": false, 00:15:41.590 "nvme_io": false, 00:15:41.590 "nvme_io_md": false, 00:15:41.590 "write_zeroes": true, 00:15:41.590 "zcopy": false, 00:15:41.590 "get_zone_info": false, 00:15:41.590 "zone_management": false, 00:15:41.590 "zone_append": false, 00:15:41.590 "compare": false, 00:15:41.590 "compare_and_write": false, 00:15:41.590 "abort": false, 00:15:41.590 "seek_hole": false, 00:15:41.590 "seek_data": false, 00:15:41.590 "copy": false, 00:15:41.590 "nvme_iov_md": false 00:15:41.590 }, 00:15:41.590 "memory_domains": [ 00:15:41.590 { 00:15:41.590 "dma_device_id": "system", 00:15:41.590 "dma_device_type": 1 00:15:41.590 }, 00:15:41.591 { 00:15:41.591 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:41.591 "dma_device_type": 2 00:15:41.591 }, 00:15:41.591 { 00:15:41.591 "dma_device_id": "system", 00:15:41.591 "dma_device_type": 1 00:15:41.591 }, 00:15:41.591 { 00:15:41.591 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:41.591 "dma_device_type": 2 00:15:41.591 } 00:15:41.591 ], 00:15:41.591 "driver_specific": { 00:15:41.591 "raid": { 00:15:41.591 "uuid": "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369", 00:15:41.591 "strip_size_kb": 0, 00:15:41.591 "state": "online", 00:15:41.591 "raid_level": "raid1", 00:15:41.591 "superblock": true, 00:15:41.591 "num_base_bdevs": 2, 00:15:41.591 "num_base_bdevs_discovered": 2, 00:15:41.591 "num_base_bdevs_operational": 2, 00:15:41.591 "base_bdevs_list": [ 00:15:41.591 { 00:15:41.591 "name": "pt1", 00:15:41.591 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:41.591 "is_configured": true, 00:15:41.591 "data_offset": 256, 00:15:41.591 "data_size": 7936 00:15:41.591 }, 00:15:41.591 { 00:15:41.591 "name": "pt2", 00:15:41.591 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:41.591 "is_configured": true, 00:15:41.591 "data_offset": 256, 00:15:41.591 "data_size": 7936 00:15:41.591 } 00:15:41.591 ] 00:15:41.591 } 00:15:41.591 } 00:15:41.591 }' 00:15:41.591 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:41.591 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:41.591 pt2' 00:15:41.591 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:41.851 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:41.852 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:41.852 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:41.852 05:43:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:41.852 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.852 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:41.852 05:43:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:41.852 [2024-12-07 05:43:15.064119] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369 '!=' b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369 ']' 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:41.852 [2024-12-07 05:43:15.111823] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:41.852 "name": "raid_bdev1", 00:15:41.852 "uuid": "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369", 00:15:41.852 "strip_size_kb": 0, 00:15:41.852 "state": "online", 00:15:41.852 "raid_level": "raid1", 00:15:41.852 "superblock": true, 00:15:41.852 "num_base_bdevs": 2, 00:15:41.852 "num_base_bdevs_discovered": 1, 00:15:41.852 "num_base_bdevs_operational": 1, 00:15:41.852 "base_bdevs_list": [ 00:15:41.852 { 00:15:41.852 "name": null, 00:15:41.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:41.852 "is_configured": false, 00:15:41.852 "data_offset": 0, 00:15:41.852 "data_size": 7936 00:15:41.852 }, 00:15:41.852 { 00:15:41.852 "name": "pt2", 00:15:41.852 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:41.852 "is_configured": true, 00:15:41.852 "data_offset": 256, 00:15:41.852 "data_size": 7936 00:15:41.852 } 00:15:41.852 ] 00:15:41.852 }' 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:41.852 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:42.426 [2024-12-07 05:43:15.527108] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:42.426 [2024-12-07 05:43:15.527183] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:42.426 [2024-12-07 05:43:15.527258] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:42.426 [2024-12-07 05:43:15.527318] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:42.426 [2024-12-07 05:43:15.527350] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.426 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:42.426 [2024-12-07 05:43:15.583009] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:42.426 [2024-12-07 05:43:15.583117] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:42.426 [2024-12-07 05:43:15.583149] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:15:42.426 [2024-12-07 05:43:15.583177] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:42.426 [2024-12-07 05:43:15.585079] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:42.426 [2024-12-07 05:43:15.585165] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:42.427 [2024-12-07 05:43:15.585234] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:42.427 [2024-12-07 05:43:15.585284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:42.427 [2024-12-07 05:43:15.585365] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:15:42.427 [2024-12-07 05:43:15.585394] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:42.427 [2024-12-07 05:43:15.585479] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:42.427 [2024-12-07 05:43:15.585583] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:15:42.427 [2024-12-07 05:43:15.585635] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:15:42.427 [2024-12-07 05:43:15.585733] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:42.427 pt2 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.427 "name": "raid_bdev1", 00:15:42.427 "uuid": "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369", 00:15:42.427 "strip_size_kb": 0, 00:15:42.427 "state": "online", 00:15:42.427 "raid_level": "raid1", 00:15:42.427 "superblock": true, 00:15:42.427 "num_base_bdevs": 2, 00:15:42.427 "num_base_bdevs_discovered": 1, 00:15:42.427 "num_base_bdevs_operational": 1, 00:15:42.427 "base_bdevs_list": [ 00:15:42.427 { 00:15:42.427 "name": null, 00:15:42.427 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.427 "is_configured": false, 00:15:42.427 "data_offset": 256, 00:15:42.427 "data_size": 7936 00:15:42.427 }, 00:15:42.427 { 00:15:42.427 "name": "pt2", 00:15:42.427 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:42.427 "is_configured": true, 00:15:42.427 "data_offset": 256, 00:15:42.427 "data_size": 7936 00:15:42.427 } 00:15:42.427 ] 00:15:42.427 }' 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.427 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:42.688 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:42.688 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.689 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:42.689 [2024-12-07 05:43:15.970371] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:42.689 [2024-12-07 05:43:15.970466] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:42.689 [2024-12-07 05:43:15.970553] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:42.689 [2024-12-07 05:43:15.970611] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:42.689 [2024-12-07 05:43:15.970700] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:15:42.689 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.689 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.689 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.689 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:42.689 05:43:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:42.689 05:43:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:42.689 [2024-12-07 05:43:16.034284] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:42.689 [2024-12-07 05:43:16.034421] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:42.689 [2024-12-07 05:43:16.034459] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:15:42.689 [2024-12-07 05:43:16.034498] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:42.689 [2024-12-07 05:43:16.036386] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:42.689 [2024-12-07 05:43:16.036453] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:42.689 [2024-12-07 05:43:16.036539] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:42.689 [2024-12-07 05:43:16.036596] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:42.689 [2024-12-07 05:43:16.036773] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:42.689 [2024-12-07 05:43:16.036842] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:42.689 [2024-12-07 05:43:16.036885] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:15:42.689 [2024-12-07 05:43:16.036951] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:42.689 [2024-12-07 05:43:16.037037] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:15:42.689 [2024-12-07 05:43:16.037079] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:42.689 [2024-12-07 05:43:16.037158] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:42.689 [2024-12-07 05:43:16.037276] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:15:42.689 [2024-12-07 05:43:16.037317] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:15:42.689 [2024-12-07 05:43:16.037434] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:42.689 pt1 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:42.689 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:42.949 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.949 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.949 "name": "raid_bdev1", 00:15:42.949 "uuid": "b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369", 00:15:42.949 "strip_size_kb": 0, 00:15:42.949 "state": "online", 00:15:42.949 "raid_level": "raid1", 00:15:42.949 "superblock": true, 00:15:42.949 "num_base_bdevs": 2, 00:15:42.949 "num_base_bdevs_discovered": 1, 00:15:42.949 "num_base_bdevs_operational": 1, 00:15:42.949 "base_bdevs_list": [ 00:15:42.949 { 00:15:42.949 "name": null, 00:15:42.949 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.949 "is_configured": false, 00:15:42.949 "data_offset": 256, 00:15:42.949 "data_size": 7936 00:15:42.949 }, 00:15:42.949 { 00:15:42.949 "name": "pt2", 00:15:42.949 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:42.949 "is_configured": true, 00:15:42.949 "data_offset": 256, 00:15:42.949 "data_size": 7936 00:15:42.949 } 00:15:42.949 ] 00:15:42.949 }' 00:15:42.949 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.949 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:43.209 [2024-12-07 05:43:16.461754] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369 '!=' b9d8c0cb-bd6b-4bd4-a3af-3a4de22ce369 ']' 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 97357 00:15:43.209 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # '[' -z 97357 ']' 00:15:43.210 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@958 -- # kill -0 97357 00:15:43.210 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # uname 00:15:43.210 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:43.210 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97357 00:15:43.210 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:43.210 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:43.210 killing process with pid 97357 00:15:43.210 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97357' 00:15:43.210 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@973 -- # kill 97357 00:15:43.210 [2024-12-07 05:43:16.533567] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:43.210 [2024-12-07 05:43:16.533641] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:43.210 [2024-12-07 05:43:16.533684] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:43.210 [2024-12-07 05:43:16.533693] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:15:43.210 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@978 -- # wait 97357 00:15:43.210 [2024-12-07 05:43:16.557332] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:43.470 05:43:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:15:43.470 00:15:43.470 real 0m4.648s 00:15:43.470 user 0m7.593s 00:15:43.470 sys 0m0.974s 00:15:43.470 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:43.470 05:43:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:43.470 ************************************ 00:15:43.470 END TEST raid_superblock_test_md_separate 00:15:43.470 ************************************ 00:15:43.470 05:43:16 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:15:43.470 05:43:16 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:15:43.470 05:43:16 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:43.470 05:43:16 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:43.470 05:43:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:43.730 ************************************ 00:15:43.730 START TEST raid_rebuild_test_sb_md_separate 00:15:43.730 ************************************ 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=97663 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 97663 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 97663 ']' 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:43.730 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:43.730 05:43:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:43.730 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:43.730 Zero copy mechanism will not be used. 00:15:43.730 [2024-12-07 05:43:16.936139] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:15:43.730 [2024-12-07 05:43:16.936319] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97663 ] 00:15:43.730 [2024-12-07 05:43:17.091811] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:43.991 [2024-12-07 05:43:17.116661] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:43.991 [2024-12-07 05:43:17.159364] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:43.991 [2024-12-07 05:43:17.159474] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:44.561 BaseBdev1_malloc 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:44.561 [2024-12-07 05:43:17.783315] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:44.561 [2024-12-07 05:43:17.783476] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:44.561 [2024-12-07 05:43:17.783522] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:44.561 [2024-12-07 05:43:17.783554] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:44.561 [2024-12-07 05:43:17.785436] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:44.561 [2024-12-07 05:43:17.785511] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:44.561 BaseBdev1 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:44.561 BaseBdev2_malloc 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:44.561 [2024-12-07 05:43:17.812309] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:44.561 [2024-12-07 05:43:17.812424] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:44.561 [2024-12-07 05:43:17.812464] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:44.561 [2024-12-07 05:43:17.812492] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:44.561 [2024-12-07 05:43:17.814385] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:44.561 [2024-12-07 05:43:17.814481] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:44.561 BaseBdev2 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:44.561 spare_malloc 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.561 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:44.561 spare_delay 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:44.562 [2024-12-07 05:43:17.860481] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:44.562 [2024-12-07 05:43:17.860607] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:44.562 [2024-12-07 05:43:17.860656] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:44.562 [2024-12-07 05:43:17.860688] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:44.562 [2024-12-07 05:43:17.862594] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:44.562 [2024-12-07 05:43:17.862707] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:44.562 spare 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:44.562 [2024-12-07 05:43:17.872501] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:44.562 [2024-12-07 05:43:17.874380] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:44.562 [2024-12-07 05:43:17.874596] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:44.562 [2024-12-07 05:43:17.874633] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:44.562 [2024-12-07 05:43:17.874786] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:44.562 [2024-12-07 05:43:17.874927] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:44.562 [2024-12-07 05:43:17.874976] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:44.562 [2024-12-07 05:43:17.875093] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:44.562 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.821 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:44.821 "name": "raid_bdev1", 00:15:44.821 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:44.821 "strip_size_kb": 0, 00:15:44.821 "state": "online", 00:15:44.821 "raid_level": "raid1", 00:15:44.821 "superblock": true, 00:15:44.821 "num_base_bdevs": 2, 00:15:44.821 "num_base_bdevs_discovered": 2, 00:15:44.821 "num_base_bdevs_operational": 2, 00:15:44.821 "base_bdevs_list": [ 00:15:44.821 { 00:15:44.821 "name": "BaseBdev1", 00:15:44.821 "uuid": "0ee7afea-7cb8-5f7b-a755-a360fe3991d4", 00:15:44.821 "is_configured": true, 00:15:44.821 "data_offset": 256, 00:15:44.821 "data_size": 7936 00:15:44.821 }, 00:15:44.821 { 00:15:44.821 "name": "BaseBdev2", 00:15:44.821 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:44.821 "is_configured": true, 00:15:44.821 "data_offset": 256, 00:15:44.821 "data_size": 7936 00:15:44.821 } 00:15:44.821 ] 00:15:44.821 }' 00:15:44.821 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:44.821 05:43:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:45.081 [2024-12-07 05:43:18.335971] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:45.081 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:45.340 [2024-12-07 05:43:18.559384] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:45.340 /dev/nbd0 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:45.340 1+0 records in 00:15:45.340 1+0 records out 00:15:45.340 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000563973 s, 7.3 MB/s 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:15:45.340 05:43:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:15:45.910 7936+0 records in 00:15:45.910 7936+0 records out 00:15:45.910 32505856 bytes (33 MB, 31 MiB) copied, 0.58196 s, 55.9 MB/s 00:15:45.910 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:45.910 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:45.910 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:45.910 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:45.910 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:15:45.910 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:45.910 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:46.170 [2024-12-07 05:43:19.427891] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.170 [2024-12-07 05:43:19.443954] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:46.170 "name": "raid_bdev1", 00:15:46.170 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:46.170 "strip_size_kb": 0, 00:15:46.170 "state": "online", 00:15:46.170 "raid_level": "raid1", 00:15:46.170 "superblock": true, 00:15:46.170 "num_base_bdevs": 2, 00:15:46.170 "num_base_bdevs_discovered": 1, 00:15:46.170 "num_base_bdevs_operational": 1, 00:15:46.170 "base_bdevs_list": [ 00:15:46.170 { 00:15:46.170 "name": null, 00:15:46.170 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:46.170 "is_configured": false, 00:15:46.170 "data_offset": 0, 00:15:46.170 "data_size": 7936 00:15:46.170 }, 00:15:46.170 { 00:15:46.170 "name": "BaseBdev2", 00:15:46.170 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:46.170 "is_configured": true, 00:15:46.170 "data_offset": 256, 00:15:46.170 "data_size": 7936 00:15:46.170 } 00:15:46.170 ] 00:15:46.170 }' 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:46.170 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.739 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:46.739 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.739 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.739 [2024-12-07 05:43:19.875280] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:46.739 [2024-12-07 05:43:19.879677] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:15:46.739 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.739 05:43:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:46.739 [2024-12-07 05:43:19.881904] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:47.677 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:47.678 "name": "raid_bdev1", 00:15:47.678 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:47.678 "strip_size_kb": 0, 00:15:47.678 "state": "online", 00:15:47.678 "raid_level": "raid1", 00:15:47.678 "superblock": true, 00:15:47.678 "num_base_bdevs": 2, 00:15:47.678 "num_base_bdevs_discovered": 2, 00:15:47.678 "num_base_bdevs_operational": 2, 00:15:47.678 "process": { 00:15:47.678 "type": "rebuild", 00:15:47.678 "target": "spare", 00:15:47.678 "progress": { 00:15:47.678 "blocks": 2560, 00:15:47.678 "percent": 32 00:15:47.678 } 00:15:47.678 }, 00:15:47.678 "base_bdevs_list": [ 00:15:47.678 { 00:15:47.678 "name": "spare", 00:15:47.678 "uuid": "216b9bcc-1b96-5dc5-9de0-2c3eabcacdba", 00:15:47.678 "is_configured": true, 00:15:47.678 "data_offset": 256, 00:15:47.678 "data_size": 7936 00:15:47.678 }, 00:15:47.678 { 00:15:47.678 "name": "BaseBdev2", 00:15:47.678 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:47.678 "is_configured": true, 00:15:47.678 "data_offset": 256, 00:15:47.678 "data_size": 7936 00:15:47.678 } 00:15:47.678 ] 00:15:47.678 }' 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:47.678 05:43:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:47.678 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:47.678 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:47.678 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.678 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.678 [2024-12-07 05:43:21.022869] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:47.937 [2024-12-07 05:43:21.090747] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:47.937 [2024-12-07 05:43:21.090813] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:47.937 [2024-12-07 05:43:21.090836] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:47.937 [2024-12-07 05:43:21.090854] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.937 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:47.937 "name": "raid_bdev1", 00:15:47.938 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:47.938 "strip_size_kb": 0, 00:15:47.938 "state": "online", 00:15:47.938 "raid_level": "raid1", 00:15:47.938 "superblock": true, 00:15:47.938 "num_base_bdevs": 2, 00:15:47.938 "num_base_bdevs_discovered": 1, 00:15:47.938 "num_base_bdevs_operational": 1, 00:15:47.938 "base_bdevs_list": [ 00:15:47.938 { 00:15:47.938 "name": null, 00:15:47.938 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:47.938 "is_configured": false, 00:15:47.938 "data_offset": 0, 00:15:47.938 "data_size": 7936 00:15:47.938 }, 00:15:47.938 { 00:15:47.938 "name": "BaseBdev2", 00:15:47.938 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:47.938 "is_configured": true, 00:15:47.938 "data_offset": 256, 00:15:47.938 "data_size": 7936 00:15:47.938 } 00:15:47.938 ] 00:15:47.938 }' 00:15:47.938 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:47.938 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.197 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:48.197 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:48.197 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:48.197 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:48.197 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:48.197 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.197 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.197 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.197 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:48.197 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.456 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:48.456 "name": "raid_bdev1", 00:15:48.456 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:48.456 "strip_size_kb": 0, 00:15:48.456 "state": "online", 00:15:48.456 "raid_level": "raid1", 00:15:48.456 "superblock": true, 00:15:48.456 "num_base_bdevs": 2, 00:15:48.456 "num_base_bdevs_discovered": 1, 00:15:48.456 "num_base_bdevs_operational": 1, 00:15:48.456 "base_bdevs_list": [ 00:15:48.456 { 00:15:48.456 "name": null, 00:15:48.456 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.456 "is_configured": false, 00:15:48.456 "data_offset": 0, 00:15:48.456 "data_size": 7936 00:15:48.456 }, 00:15:48.456 { 00:15:48.456 "name": "BaseBdev2", 00:15:48.456 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:48.456 "is_configured": true, 00:15:48.456 "data_offset": 256, 00:15:48.456 "data_size": 7936 00:15:48.456 } 00:15:48.456 ] 00:15:48.456 }' 00:15:48.456 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:48.456 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:48.456 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:48.456 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:48.456 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:48.456 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.456 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.456 [2024-12-07 05:43:21.687814] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:48.456 [2024-12-07 05:43:21.692232] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:15:48.456 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.456 05:43:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:48.456 [2024-12-07 05:43:21.694461] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:49.410 "name": "raid_bdev1", 00:15:49.410 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:49.410 "strip_size_kb": 0, 00:15:49.410 "state": "online", 00:15:49.410 "raid_level": "raid1", 00:15:49.410 "superblock": true, 00:15:49.410 "num_base_bdevs": 2, 00:15:49.410 "num_base_bdevs_discovered": 2, 00:15:49.410 "num_base_bdevs_operational": 2, 00:15:49.410 "process": { 00:15:49.410 "type": "rebuild", 00:15:49.410 "target": "spare", 00:15:49.410 "progress": { 00:15:49.410 "blocks": 2560, 00:15:49.410 "percent": 32 00:15:49.410 } 00:15:49.410 }, 00:15:49.410 "base_bdevs_list": [ 00:15:49.410 { 00:15:49.410 "name": "spare", 00:15:49.410 "uuid": "216b9bcc-1b96-5dc5-9de0-2c3eabcacdba", 00:15:49.410 "is_configured": true, 00:15:49.410 "data_offset": 256, 00:15:49.410 "data_size": 7936 00:15:49.410 }, 00:15:49.410 { 00:15:49.410 "name": "BaseBdev2", 00:15:49.410 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:49.410 "is_configured": true, 00:15:49.410 "data_offset": 256, 00:15:49.410 "data_size": 7936 00:15:49.410 } 00:15:49.410 ] 00:15:49.410 }' 00:15:49.410 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:49.670 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:49.670 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:49.670 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:49.670 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:49.670 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:49.670 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=579 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:49.671 "name": "raid_bdev1", 00:15:49.671 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:49.671 "strip_size_kb": 0, 00:15:49.671 "state": "online", 00:15:49.671 "raid_level": "raid1", 00:15:49.671 "superblock": true, 00:15:49.671 "num_base_bdevs": 2, 00:15:49.671 "num_base_bdevs_discovered": 2, 00:15:49.671 "num_base_bdevs_operational": 2, 00:15:49.671 "process": { 00:15:49.671 "type": "rebuild", 00:15:49.671 "target": "spare", 00:15:49.671 "progress": { 00:15:49.671 "blocks": 2816, 00:15:49.671 "percent": 35 00:15:49.671 } 00:15:49.671 }, 00:15:49.671 "base_bdevs_list": [ 00:15:49.671 { 00:15:49.671 "name": "spare", 00:15:49.671 "uuid": "216b9bcc-1b96-5dc5-9de0-2c3eabcacdba", 00:15:49.671 "is_configured": true, 00:15:49.671 "data_offset": 256, 00:15:49.671 "data_size": 7936 00:15:49.671 }, 00:15:49.671 { 00:15:49.671 "name": "BaseBdev2", 00:15:49.671 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:49.671 "is_configured": true, 00:15:49.671 "data_offset": 256, 00:15:49.671 "data_size": 7936 00:15:49.671 } 00:15:49.671 ] 00:15:49.671 }' 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:49.671 05:43:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:50.607 05:43:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:50.607 05:43:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:50.607 05:43:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:50.607 05:43:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:50.607 05:43:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:50.607 05:43:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:50.866 05:43:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.866 05:43:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:50.866 05:43:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.866 05:43:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.866 05:43:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.866 05:43:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:50.866 "name": "raid_bdev1", 00:15:50.866 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:50.866 "strip_size_kb": 0, 00:15:50.866 "state": "online", 00:15:50.866 "raid_level": "raid1", 00:15:50.866 "superblock": true, 00:15:50.866 "num_base_bdevs": 2, 00:15:50.866 "num_base_bdevs_discovered": 2, 00:15:50.866 "num_base_bdevs_operational": 2, 00:15:50.866 "process": { 00:15:50.866 "type": "rebuild", 00:15:50.866 "target": "spare", 00:15:50.866 "progress": { 00:15:50.866 "blocks": 5632, 00:15:50.866 "percent": 70 00:15:50.866 } 00:15:50.866 }, 00:15:50.866 "base_bdevs_list": [ 00:15:50.866 { 00:15:50.866 "name": "spare", 00:15:50.866 "uuid": "216b9bcc-1b96-5dc5-9de0-2c3eabcacdba", 00:15:50.866 "is_configured": true, 00:15:50.866 "data_offset": 256, 00:15:50.866 "data_size": 7936 00:15:50.866 }, 00:15:50.866 { 00:15:50.866 "name": "BaseBdev2", 00:15:50.866 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:50.866 "is_configured": true, 00:15:50.866 "data_offset": 256, 00:15:50.866 "data_size": 7936 00:15:50.866 } 00:15:50.867 ] 00:15:50.867 }' 00:15:50.867 05:43:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:50.867 05:43:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:50.867 05:43:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:50.867 05:43:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:50.867 05:43:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:51.805 [2024-12-07 05:43:24.815785] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:51.805 [2024-12-07 05:43:24.815973] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:51.805 [2024-12-07 05:43:24.816122] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.805 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:51.805 "name": "raid_bdev1", 00:15:51.805 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:51.805 "strip_size_kb": 0, 00:15:51.806 "state": "online", 00:15:51.806 "raid_level": "raid1", 00:15:51.806 "superblock": true, 00:15:51.806 "num_base_bdevs": 2, 00:15:51.806 "num_base_bdevs_discovered": 2, 00:15:51.806 "num_base_bdevs_operational": 2, 00:15:51.806 "base_bdevs_list": [ 00:15:51.806 { 00:15:51.806 "name": "spare", 00:15:51.806 "uuid": "216b9bcc-1b96-5dc5-9de0-2c3eabcacdba", 00:15:51.806 "is_configured": true, 00:15:51.806 "data_offset": 256, 00:15:51.806 "data_size": 7936 00:15:51.806 }, 00:15:51.806 { 00:15:51.806 "name": "BaseBdev2", 00:15:51.806 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:51.806 "is_configured": true, 00:15:51.806 "data_offset": 256, 00:15:51.806 "data_size": 7936 00:15:51.806 } 00:15:51.806 ] 00:15:51.806 }' 00:15:51.806 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:52.064 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:52.065 "name": "raid_bdev1", 00:15:52.065 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:52.065 "strip_size_kb": 0, 00:15:52.065 "state": "online", 00:15:52.065 "raid_level": "raid1", 00:15:52.065 "superblock": true, 00:15:52.065 "num_base_bdevs": 2, 00:15:52.065 "num_base_bdevs_discovered": 2, 00:15:52.065 "num_base_bdevs_operational": 2, 00:15:52.065 "base_bdevs_list": [ 00:15:52.065 { 00:15:52.065 "name": "spare", 00:15:52.065 "uuid": "216b9bcc-1b96-5dc5-9de0-2c3eabcacdba", 00:15:52.065 "is_configured": true, 00:15:52.065 "data_offset": 256, 00:15:52.065 "data_size": 7936 00:15:52.065 }, 00:15:52.065 { 00:15:52.065 "name": "BaseBdev2", 00:15:52.065 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:52.065 "is_configured": true, 00:15:52.065 "data_offset": 256, 00:15:52.065 "data_size": 7936 00:15:52.065 } 00:15:52.065 ] 00:15:52.065 }' 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.065 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.325 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:52.325 "name": "raid_bdev1", 00:15:52.325 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:52.325 "strip_size_kb": 0, 00:15:52.325 "state": "online", 00:15:52.325 "raid_level": "raid1", 00:15:52.325 "superblock": true, 00:15:52.325 "num_base_bdevs": 2, 00:15:52.325 "num_base_bdevs_discovered": 2, 00:15:52.325 "num_base_bdevs_operational": 2, 00:15:52.325 "base_bdevs_list": [ 00:15:52.325 { 00:15:52.325 "name": "spare", 00:15:52.325 "uuid": "216b9bcc-1b96-5dc5-9de0-2c3eabcacdba", 00:15:52.325 "is_configured": true, 00:15:52.325 "data_offset": 256, 00:15:52.325 "data_size": 7936 00:15:52.325 }, 00:15:52.325 { 00:15:52.325 "name": "BaseBdev2", 00:15:52.325 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:52.325 "is_configured": true, 00:15:52.325 "data_offset": 256, 00:15:52.325 "data_size": 7936 00:15:52.325 } 00:15:52.325 ] 00:15:52.325 }' 00:15:52.325 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:52.325 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.584 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:52.584 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.584 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.585 [2024-12-07 05:43:25.820035] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:52.585 [2024-12-07 05:43:25.820114] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:52.585 [2024-12-07 05:43:25.820260] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:52.585 [2024-12-07 05:43:25.820374] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:52.585 [2024-12-07 05:43:25.820424] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:52.585 05:43:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:52.845 /dev/nbd0 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:52.845 1+0 records in 00:15:52.845 1+0 records out 00:15:52.845 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000525089 s, 7.8 MB/s 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:52.845 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:53.106 /dev/nbd1 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:53.106 1+0 records in 00:15:53.106 1+0 records out 00:15:53.106 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000371868 s, 11.0 MB/s 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:53.106 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:53.366 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:53.366 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:53.366 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:53.366 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:53.366 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:53.366 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:53.366 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:15:53.366 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:15:53.366 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:53.366 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.626 [2024-12-07 05:43:26.868569] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:53.626 [2024-12-07 05:43:26.868644] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:53.626 [2024-12-07 05:43:26.868668] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:53.626 [2024-12-07 05:43:26.868683] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:53.626 [2024-12-07 05:43:26.870971] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:53.626 spare 00:15:53.626 [2024-12-07 05:43:26.871057] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:53.626 [2024-12-07 05:43:26.871128] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:53.626 [2024-12-07 05:43:26.871173] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:53.626 [2024-12-07 05:43:26.871296] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.626 [2024-12-07 05:43:26.971190] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:53.626 [2024-12-07 05:43:26.971254] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:53.626 [2024-12-07 05:43:26.971397] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:15:53.626 [2024-12-07 05:43:26.971565] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:53.626 [2024-12-07 05:43:26.971613] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:53.626 [2024-12-07 05:43:26.971783] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.626 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.886 05:43:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.886 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.886 "name": "raid_bdev1", 00:15:53.886 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:53.886 "strip_size_kb": 0, 00:15:53.886 "state": "online", 00:15:53.886 "raid_level": "raid1", 00:15:53.886 "superblock": true, 00:15:53.886 "num_base_bdevs": 2, 00:15:53.887 "num_base_bdevs_discovered": 2, 00:15:53.887 "num_base_bdevs_operational": 2, 00:15:53.887 "base_bdevs_list": [ 00:15:53.887 { 00:15:53.887 "name": "spare", 00:15:53.887 "uuid": "216b9bcc-1b96-5dc5-9de0-2c3eabcacdba", 00:15:53.887 "is_configured": true, 00:15:53.887 "data_offset": 256, 00:15:53.887 "data_size": 7936 00:15:53.887 }, 00:15:53.887 { 00:15:53.887 "name": "BaseBdev2", 00:15:53.887 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:53.887 "is_configured": true, 00:15:53.887 "data_offset": 256, 00:15:53.887 "data_size": 7936 00:15:53.887 } 00:15:53.887 ] 00:15:53.887 }' 00:15:53.887 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.887 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.146 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:54.146 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:54.146 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:54.146 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:54.146 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:54.146 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:54.146 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.146 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.146 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.146 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.146 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:54.146 "name": "raid_bdev1", 00:15:54.146 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:54.146 "strip_size_kb": 0, 00:15:54.146 "state": "online", 00:15:54.146 "raid_level": "raid1", 00:15:54.146 "superblock": true, 00:15:54.146 "num_base_bdevs": 2, 00:15:54.146 "num_base_bdevs_discovered": 2, 00:15:54.146 "num_base_bdevs_operational": 2, 00:15:54.146 "base_bdevs_list": [ 00:15:54.146 { 00:15:54.146 "name": "spare", 00:15:54.146 "uuid": "216b9bcc-1b96-5dc5-9de0-2c3eabcacdba", 00:15:54.146 "is_configured": true, 00:15:54.146 "data_offset": 256, 00:15:54.147 "data_size": 7936 00:15:54.147 }, 00:15:54.147 { 00:15:54.147 "name": "BaseBdev2", 00:15:54.147 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:54.147 "is_configured": true, 00:15:54.147 "data_offset": 256, 00:15:54.147 "data_size": 7936 00:15:54.147 } 00:15:54.147 ] 00:15:54.147 }' 00:15:54.147 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:54.406 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:54.406 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:54.406 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.407 [2024-12-07 05:43:27.599414] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:54.407 "name": "raid_bdev1", 00:15:54.407 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:54.407 "strip_size_kb": 0, 00:15:54.407 "state": "online", 00:15:54.407 "raid_level": "raid1", 00:15:54.407 "superblock": true, 00:15:54.407 "num_base_bdevs": 2, 00:15:54.407 "num_base_bdevs_discovered": 1, 00:15:54.407 "num_base_bdevs_operational": 1, 00:15:54.407 "base_bdevs_list": [ 00:15:54.407 { 00:15:54.407 "name": null, 00:15:54.407 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:54.407 "is_configured": false, 00:15:54.407 "data_offset": 0, 00:15:54.407 "data_size": 7936 00:15:54.407 }, 00:15:54.407 { 00:15:54.407 "name": "BaseBdev2", 00:15:54.407 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:54.407 "is_configured": true, 00:15:54.407 "data_offset": 256, 00:15:54.407 "data_size": 7936 00:15:54.407 } 00:15:54.407 ] 00:15:54.407 }' 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:54.407 05:43:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.977 05:43:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:54.977 05:43:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.977 05:43:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.977 [2024-12-07 05:43:28.054762] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:54.977 [2024-12-07 05:43:28.055039] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:54.977 [2024-12-07 05:43:28.055100] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:54.977 [2024-12-07 05:43:28.055172] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:54.977 [2024-12-07 05:43:28.059587] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:15:54.977 05:43:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.977 05:43:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:54.977 [2024-12-07 05:43:28.061790] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:55.917 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:55.917 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:55.917 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:55.917 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:55.917 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:55.917 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.917 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:55.918 "name": "raid_bdev1", 00:15:55.918 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:55.918 "strip_size_kb": 0, 00:15:55.918 "state": "online", 00:15:55.918 "raid_level": "raid1", 00:15:55.918 "superblock": true, 00:15:55.918 "num_base_bdevs": 2, 00:15:55.918 "num_base_bdevs_discovered": 2, 00:15:55.918 "num_base_bdevs_operational": 2, 00:15:55.918 "process": { 00:15:55.918 "type": "rebuild", 00:15:55.918 "target": "spare", 00:15:55.918 "progress": { 00:15:55.918 "blocks": 2560, 00:15:55.918 "percent": 32 00:15:55.918 } 00:15:55.918 }, 00:15:55.918 "base_bdevs_list": [ 00:15:55.918 { 00:15:55.918 "name": "spare", 00:15:55.918 "uuid": "216b9bcc-1b96-5dc5-9de0-2c3eabcacdba", 00:15:55.918 "is_configured": true, 00:15:55.918 "data_offset": 256, 00:15:55.918 "data_size": 7936 00:15:55.918 }, 00:15:55.918 { 00:15:55.918 "name": "BaseBdev2", 00:15:55.918 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:55.918 "is_configured": true, 00:15:55.918 "data_offset": 256, 00:15:55.918 "data_size": 7936 00:15:55.918 } 00:15:55.918 ] 00:15:55.918 }' 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.918 [2024-12-07 05:43:29.198926] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:55.918 [2024-12-07 05:43:29.269870] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:55.918 [2024-12-07 05:43:29.270408] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:55.918 [2024-12-07 05:43:29.270491] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:55.918 [2024-12-07 05:43:29.270517] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:55.918 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:56.178 "name": "raid_bdev1", 00:15:56.178 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:56.178 "strip_size_kb": 0, 00:15:56.178 "state": "online", 00:15:56.178 "raid_level": "raid1", 00:15:56.178 "superblock": true, 00:15:56.178 "num_base_bdevs": 2, 00:15:56.178 "num_base_bdevs_discovered": 1, 00:15:56.178 "num_base_bdevs_operational": 1, 00:15:56.178 "base_bdevs_list": [ 00:15:56.178 { 00:15:56.178 "name": null, 00:15:56.178 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:56.178 "is_configured": false, 00:15:56.178 "data_offset": 0, 00:15:56.178 "data_size": 7936 00:15:56.178 }, 00:15:56.178 { 00:15:56.178 "name": "BaseBdev2", 00:15:56.178 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:56.178 "is_configured": true, 00:15:56.178 "data_offset": 256, 00:15:56.178 "data_size": 7936 00:15:56.178 } 00:15:56.178 ] 00:15:56.178 }' 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:56.178 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:56.438 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:56.438 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.438 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:56.438 [2024-12-07 05:43:29.751537] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:56.438 [2024-12-07 05:43:29.751901] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:56.438 [2024-12-07 05:43:29.751982] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:15:56.438 [2024-12-07 05:43:29.752066] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:56.438 [2024-12-07 05:43:29.752390] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:56.438 [2024-12-07 05:43:29.752505] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:56.438 [2024-12-07 05:43:29.752660] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:56.438 [2024-12-07 05:43:29.752704] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:56.438 [2024-12-07 05:43:29.752752] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:56.438 [2024-12-07 05:43:29.752881] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:56.438 [2024-12-07 05:43:29.757142] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:15:56.438 spare 00:15:56.438 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.438 05:43:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:56.438 [2024-12-07 05:43:29.759349] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:57.818 "name": "raid_bdev1", 00:15:57.818 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:57.818 "strip_size_kb": 0, 00:15:57.818 "state": "online", 00:15:57.818 "raid_level": "raid1", 00:15:57.818 "superblock": true, 00:15:57.818 "num_base_bdevs": 2, 00:15:57.818 "num_base_bdevs_discovered": 2, 00:15:57.818 "num_base_bdevs_operational": 2, 00:15:57.818 "process": { 00:15:57.818 "type": "rebuild", 00:15:57.818 "target": "spare", 00:15:57.818 "progress": { 00:15:57.818 "blocks": 2560, 00:15:57.818 "percent": 32 00:15:57.818 } 00:15:57.818 }, 00:15:57.818 "base_bdevs_list": [ 00:15:57.818 { 00:15:57.818 "name": "spare", 00:15:57.818 "uuid": "216b9bcc-1b96-5dc5-9de0-2c3eabcacdba", 00:15:57.818 "is_configured": true, 00:15:57.818 "data_offset": 256, 00:15:57.818 "data_size": 7936 00:15:57.818 }, 00:15:57.818 { 00:15:57.818 "name": "BaseBdev2", 00:15:57.818 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:57.818 "is_configured": true, 00:15:57.818 "data_offset": 256, 00:15:57.818 "data_size": 7936 00:15:57.818 } 00:15:57.818 ] 00:15:57.818 }' 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.818 [2024-12-07 05:43:30.902966] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:57.818 [2024-12-07 05:43:30.967557] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:57.818 [2024-12-07 05:43:30.968100] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:57.818 [2024-12-07 05:43:30.968162] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:57.818 [2024-12-07 05:43:30.968191] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.818 05:43:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.818 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.818 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:57.818 "name": "raid_bdev1", 00:15:57.818 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:57.818 "strip_size_kb": 0, 00:15:57.818 "state": "online", 00:15:57.818 "raid_level": "raid1", 00:15:57.818 "superblock": true, 00:15:57.818 "num_base_bdevs": 2, 00:15:57.818 "num_base_bdevs_discovered": 1, 00:15:57.818 "num_base_bdevs_operational": 1, 00:15:57.818 "base_bdevs_list": [ 00:15:57.818 { 00:15:57.818 "name": null, 00:15:57.818 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:57.818 "is_configured": false, 00:15:57.818 "data_offset": 0, 00:15:57.818 "data_size": 7936 00:15:57.818 }, 00:15:57.818 { 00:15:57.818 "name": "BaseBdev2", 00:15:57.818 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:57.818 "is_configured": true, 00:15:57.818 "data_offset": 256, 00:15:57.818 "data_size": 7936 00:15:57.818 } 00:15:57.818 ] 00:15:57.818 }' 00:15:57.818 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:57.818 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.078 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:58.337 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:58.337 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:58.337 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:58.338 "name": "raid_bdev1", 00:15:58.338 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:58.338 "strip_size_kb": 0, 00:15:58.338 "state": "online", 00:15:58.338 "raid_level": "raid1", 00:15:58.338 "superblock": true, 00:15:58.338 "num_base_bdevs": 2, 00:15:58.338 "num_base_bdevs_discovered": 1, 00:15:58.338 "num_base_bdevs_operational": 1, 00:15:58.338 "base_bdevs_list": [ 00:15:58.338 { 00:15:58.338 "name": null, 00:15:58.338 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:58.338 "is_configured": false, 00:15:58.338 "data_offset": 0, 00:15:58.338 "data_size": 7936 00:15:58.338 }, 00:15:58.338 { 00:15:58.338 "name": "BaseBdev2", 00:15:58.338 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:58.338 "is_configured": true, 00:15:58.338 "data_offset": 256, 00:15:58.338 "data_size": 7936 00:15:58.338 } 00:15:58.338 ] 00:15:58.338 }' 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.338 [2024-12-07 05:43:31.616798] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:58.338 [2024-12-07 05:43:31.616921] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:58.338 [2024-12-07 05:43:31.616963] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:15:58.338 [2024-12-07 05:43:31.617003] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:58.338 [2024-12-07 05:43:31.617251] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:58.338 [2024-12-07 05:43:31.617305] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:58.338 [2024-12-07 05:43:31.617386] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:58.338 [2024-12-07 05:43:31.617435] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:58.338 [2024-12-07 05:43:31.617471] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:58.338 [2024-12-07 05:43:31.617541] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:58.338 BaseBdev1 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.338 05:43:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.277 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.537 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.537 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:59.537 "name": "raid_bdev1", 00:15:59.537 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:59.537 "strip_size_kb": 0, 00:15:59.537 "state": "online", 00:15:59.537 "raid_level": "raid1", 00:15:59.537 "superblock": true, 00:15:59.537 "num_base_bdevs": 2, 00:15:59.537 "num_base_bdevs_discovered": 1, 00:15:59.537 "num_base_bdevs_operational": 1, 00:15:59.537 "base_bdevs_list": [ 00:15:59.537 { 00:15:59.537 "name": null, 00:15:59.537 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:59.537 "is_configured": false, 00:15:59.537 "data_offset": 0, 00:15:59.537 "data_size": 7936 00:15:59.537 }, 00:15:59.537 { 00:15:59.537 "name": "BaseBdev2", 00:15:59.537 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:59.537 "is_configured": true, 00:15:59.537 "data_offset": 256, 00:15:59.537 "data_size": 7936 00:15:59.537 } 00:15:59.537 ] 00:15:59.537 }' 00:15:59.537 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:59.537 05:43:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:59.797 "name": "raid_bdev1", 00:15:59.797 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:15:59.797 "strip_size_kb": 0, 00:15:59.797 "state": "online", 00:15:59.797 "raid_level": "raid1", 00:15:59.797 "superblock": true, 00:15:59.797 "num_base_bdevs": 2, 00:15:59.797 "num_base_bdevs_discovered": 1, 00:15:59.797 "num_base_bdevs_operational": 1, 00:15:59.797 "base_bdevs_list": [ 00:15:59.797 { 00:15:59.797 "name": null, 00:15:59.797 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:59.797 "is_configured": false, 00:15:59.797 "data_offset": 0, 00:15:59.797 "data_size": 7936 00:15:59.797 }, 00:15:59.797 { 00:15:59.797 "name": "BaseBdev2", 00:15:59.797 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:15:59.797 "is_configured": true, 00:15:59.797 "data_offset": 256, 00:15:59.797 "data_size": 7936 00:15:59.797 } 00:15:59.797 ] 00:15:59.797 }' 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:59.797 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.057 [2024-12-07 05:43:33.210198] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:00.057 [2024-12-07 05:43:33.210417] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:00.057 [2024-12-07 05:43:33.210473] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:00.057 request: 00:16:00.057 { 00:16:00.057 "base_bdev": "BaseBdev1", 00:16:00.057 "raid_bdev": "raid_bdev1", 00:16:00.057 "method": "bdev_raid_add_base_bdev", 00:16:00.057 "req_id": 1 00:16:00.057 } 00:16:00.057 Got JSON-RPC error response 00:16:00.057 response: 00:16:00.057 { 00:16:00.057 "code": -22, 00:16:00.057 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:00.057 } 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # es=1 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:00.057 05:43:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:00.997 "name": "raid_bdev1", 00:16:00.997 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:16:00.997 "strip_size_kb": 0, 00:16:00.997 "state": "online", 00:16:00.997 "raid_level": "raid1", 00:16:00.997 "superblock": true, 00:16:00.997 "num_base_bdevs": 2, 00:16:00.997 "num_base_bdevs_discovered": 1, 00:16:00.997 "num_base_bdevs_operational": 1, 00:16:00.997 "base_bdevs_list": [ 00:16:00.997 { 00:16:00.997 "name": null, 00:16:00.997 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:00.997 "is_configured": false, 00:16:00.997 "data_offset": 0, 00:16:00.997 "data_size": 7936 00:16:00.997 }, 00:16:00.997 { 00:16:00.997 "name": "BaseBdev2", 00:16:00.997 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:16:00.997 "is_configured": true, 00:16:00.997 "data_offset": 256, 00:16:00.997 "data_size": 7936 00:16:00.997 } 00:16:00.997 ] 00:16:00.997 }' 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:00.997 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:01.567 "name": "raid_bdev1", 00:16:01.567 "uuid": "8f53242b-dca6-4e2f-940a-59eea1664e47", 00:16:01.567 "strip_size_kb": 0, 00:16:01.567 "state": "online", 00:16:01.567 "raid_level": "raid1", 00:16:01.567 "superblock": true, 00:16:01.567 "num_base_bdevs": 2, 00:16:01.567 "num_base_bdevs_discovered": 1, 00:16:01.567 "num_base_bdevs_operational": 1, 00:16:01.567 "base_bdevs_list": [ 00:16:01.567 { 00:16:01.567 "name": null, 00:16:01.567 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:01.567 "is_configured": false, 00:16:01.567 "data_offset": 0, 00:16:01.567 "data_size": 7936 00:16:01.567 }, 00:16:01.567 { 00:16:01.567 "name": "BaseBdev2", 00:16:01.567 "uuid": "e0df2606-10d6-5693-99cd-ca6d25245711", 00:16:01.567 "is_configured": true, 00:16:01.567 "data_offset": 256, 00:16:01.567 "data_size": 7936 00:16:01.567 } 00:16:01.567 ] 00:16:01.567 }' 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 97663 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 97663 ']' 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 97663 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97663 00:16:01.567 killing process with pid 97663 00:16:01.567 Received shutdown signal, test time was about 60.000000 seconds 00:16:01.567 00:16:01.567 Latency(us) 00:16:01.567 [2024-12-07T05:43:34.935Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:01.567 [2024-12-07T05:43:34.935Z] =================================================================================================================== 00:16:01.567 [2024-12-07T05:43:34.935Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97663' 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 97663 00:16:01.567 05:43:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 97663 00:16:01.567 [2024-12-07 05:43:34.797502] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:01.567 [2024-12-07 05:43:34.797616] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:01.567 [2024-12-07 05:43:34.797693] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:01.567 [2024-12-07 05:43:34.797721] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:16:01.567 [2024-12-07 05:43:34.830626] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:01.827 05:43:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:16:01.827 00:16:01.827 real 0m18.180s 00:16:01.827 user 0m24.193s 00:16:01.827 sys 0m2.398s 00:16:01.827 05:43:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:01.827 05:43:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.827 ************************************ 00:16:01.827 END TEST raid_rebuild_test_sb_md_separate 00:16:01.827 ************************************ 00:16:01.827 05:43:35 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:16:01.827 05:43:35 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:16:01.827 05:43:35 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:16:01.827 05:43:35 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:01.827 05:43:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:01.827 ************************************ 00:16:01.827 START TEST raid_state_function_test_sb_md_interleaved 00:16:01.827 ************************************ 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:01.827 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:01.828 Process raid pid: 98341 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=98341 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 98341' 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 98341 00:16:01.828 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 98341 ']' 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:01.828 05:43:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:02.086 [2024-12-07 05:43:35.193690] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:16:02.087 [2024-12-07 05:43:35.193804] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:02.087 [2024-12-07 05:43:35.349476] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:02.087 [2024-12-07 05:43:35.373958] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:02.087 [2024-12-07 05:43:35.416056] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:02.087 [2024-12-07 05:43:35.416096] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:02.654 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:02.655 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:16:02.655 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:02.655 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.655 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:02.655 [2024-12-07 05:43:36.018658] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:02.655 [2024-12-07 05:43:36.018781] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:02.655 [2024-12-07 05:43:36.018824] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:02.655 [2024-12-07 05:43:36.018849] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:02.916 "name": "Existed_Raid", 00:16:02.916 "uuid": "f6e267a8-5ce3-4745-86c0-12d6105d9403", 00:16:02.916 "strip_size_kb": 0, 00:16:02.916 "state": "configuring", 00:16:02.916 "raid_level": "raid1", 00:16:02.916 "superblock": true, 00:16:02.916 "num_base_bdevs": 2, 00:16:02.916 "num_base_bdevs_discovered": 0, 00:16:02.916 "num_base_bdevs_operational": 2, 00:16:02.916 "base_bdevs_list": [ 00:16:02.916 { 00:16:02.916 "name": "BaseBdev1", 00:16:02.916 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:02.916 "is_configured": false, 00:16:02.916 "data_offset": 0, 00:16:02.916 "data_size": 0 00:16:02.916 }, 00:16:02.916 { 00:16:02.916 "name": "BaseBdev2", 00:16:02.916 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:02.916 "is_configured": false, 00:16:02.916 "data_offset": 0, 00:16:02.916 "data_size": 0 00:16:02.916 } 00:16:02.916 ] 00:16:02.916 }' 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:02.916 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:03.176 [2024-12-07 05:43:36.461822] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:03.176 [2024-12-07 05:43:36.461904] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:03.176 [2024-12-07 05:43:36.473802] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:03.176 [2024-12-07 05:43:36.473880] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:03.176 [2024-12-07 05:43:36.473908] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:03.176 [2024-12-07 05:43:36.473942] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:03.176 [2024-12-07 05:43:36.494755] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:03.176 BaseBdev1 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:03.176 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:03.177 [ 00:16:03.177 { 00:16:03.177 "name": "BaseBdev1", 00:16:03.177 "aliases": [ 00:16:03.177 "f92752e2-29da-49c0-a5b3-5922b6d45994" 00:16:03.177 ], 00:16:03.177 "product_name": "Malloc disk", 00:16:03.177 "block_size": 4128, 00:16:03.177 "num_blocks": 8192, 00:16:03.177 "uuid": "f92752e2-29da-49c0-a5b3-5922b6d45994", 00:16:03.177 "md_size": 32, 00:16:03.177 "md_interleave": true, 00:16:03.177 "dif_type": 0, 00:16:03.177 "assigned_rate_limits": { 00:16:03.177 "rw_ios_per_sec": 0, 00:16:03.177 "rw_mbytes_per_sec": 0, 00:16:03.177 "r_mbytes_per_sec": 0, 00:16:03.177 "w_mbytes_per_sec": 0 00:16:03.177 }, 00:16:03.177 "claimed": true, 00:16:03.177 "claim_type": "exclusive_write", 00:16:03.177 "zoned": false, 00:16:03.177 "supported_io_types": { 00:16:03.177 "read": true, 00:16:03.177 "write": true, 00:16:03.177 "unmap": true, 00:16:03.177 "flush": true, 00:16:03.177 "reset": true, 00:16:03.177 "nvme_admin": false, 00:16:03.177 "nvme_io": false, 00:16:03.177 "nvme_io_md": false, 00:16:03.177 "write_zeroes": true, 00:16:03.177 "zcopy": true, 00:16:03.177 "get_zone_info": false, 00:16:03.177 "zone_management": false, 00:16:03.177 "zone_append": false, 00:16:03.177 "compare": false, 00:16:03.177 "compare_and_write": false, 00:16:03.177 "abort": true, 00:16:03.177 "seek_hole": false, 00:16:03.177 "seek_data": false, 00:16:03.177 "copy": true, 00:16:03.177 "nvme_iov_md": false 00:16:03.177 }, 00:16:03.177 "memory_domains": [ 00:16:03.177 { 00:16:03.177 "dma_device_id": "system", 00:16:03.177 "dma_device_type": 1 00:16:03.177 }, 00:16:03.177 { 00:16:03.177 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:03.177 "dma_device_type": 2 00:16:03.177 } 00:16:03.177 ], 00:16:03.177 "driver_specific": {} 00:16:03.177 } 00:16:03.177 ] 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:03.177 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.436 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:03.436 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.436 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:03.436 "name": "Existed_Raid", 00:16:03.436 "uuid": "6f76cd73-5f01-4100-9654-34aed85452ac", 00:16:03.436 "strip_size_kb": 0, 00:16:03.436 "state": "configuring", 00:16:03.436 "raid_level": "raid1", 00:16:03.436 "superblock": true, 00:16:03.436 "num_base_bdevs": 2, 00:16:03.436 "num_base_bdevs_discovered": 1, 00:16:03.436 "num_base_bdevs_operational": 2, 00:16:03.436 "base_bdevs_list": [ 00:16:03.436 { 00:16:03.436 "name": "BaseBdev1", 00:16:03.436 "uuid": "f92752e2-29da-49c0-a5b3-5922b6d45994", 00:16:03.436 "is_configured": true, 00:16:03.436 "data_offset": 256, 00:16:03.436 "data_size": 7936 00:16:03.436 }, 00:16:03.436 { 00:16:03.436 "name": "BaseBdev2", 00:16:03.436 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:03.436 "is_configured": false, 00:16:03.436 "data_offset": 0, 00:16:03.436 "data_size": 0 00:16:03.436 } 00:16:03.436 ] 00:16:03.436 }' 00:16:03.436 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:03.436 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:03.695 [2024-12-07 05:43:36.970001] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:03.695 [2024-12-07 05:43:36.970100] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:03.695 [2024-12-07 05:43:36.982034] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:03.695 [2024-12-07 05:43:36.983885] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:03.695 [2024-12-07 05:43:36.983960] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:03.695 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:03.696 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:03.696 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:03.696 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:03.696 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:03.696 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.696 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:03.696 05:43:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:03.696 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.696 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:03.696 "name": "Existed_Raid", 00:16:03.696 "uuid": "5c8b56f3-99fb-4e43-af30-10fc26306a90", 00:16:03.696 "strip_size_kb": 0, 00:16:03.696 "state": "configuring", 00:16:03.696 "raid_level": "raid1", 00:16:03.696 "superblock": true, 00:16:03.696 "num_base_bdevs": 2, 00:16:03.696 "num_base_bdevs_discovered": 1, 00:16:03.696 "num_base_bdevs_operational": 2, 00:16:03.696 "base_bdevs_list": [ 00:16:03.696 { 00:16:03.696 "name": "BaseBdev1", 00:16:03.696 "uuid": "f92752e2-29da-49c0-a5b3-5922b6d45994", 00:16:03.696 "is_configured": true, 00:16:03.696 "data_offset": 256, 00:16:03.696 "data_size": 7936 00:16:03.696 }, 00:16:03.696 { 00:16:03.696 "name": "BaseBdev2", 00:16:03.696 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:03.696 "is_configured": false, 00:16:03.696 "data_offset": 0, 00:16:03.696 "data_size": 0 00:16:03.696 } 00:16:03.696 ] 00:16:03.696 }' 00:16:03.696 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:03.696 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:04.266 [2024-12-07 05:43:37.424388] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:04.266 [2024-12-07 05:43:37.424566] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:04.266 [2024-12-07 05:43:37.424580] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:04.266 [2024-12-07 05:43:37.424708] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:04.266 [2024-12-07 05:43:37.424805] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:04.266 [2024-12-07 05:43:37.424819] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:16:04.266 [2024-12-07 05:43:37.424883] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:04.266 BaseBdev2 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:04.266 [ 00:16:04.266 { 00:16:04.266 "name": "BaseBdev2", 00:16:04.266 "aliases": [ 00:16:04.266 "d6ce9d07-2996-439c-ae06-bfe71ca55548" 00:16:04.266 ], 00:16:04.266 "product_name": "Malloc disk", 00:16:04.266 "block_size": 4128, 00:16:04.266 "num_blocks": 8192, 00:16:04.266 "uuid": "d6ce9d07-2996-439c-ae06-bfe71ca55548", 00:16:04.266 "md_size": 32, 00:16:04.266 "md_interleave": true, 00:16:04.266 "dif_type": 0, 00:16:04.266 "assigned_rate_limits": { 00:16:04.266 "rw_ios_per_sec": 0, 00:16:04.266 "rw_mbytes_per_sec": 0, 00:16:04.266 "r_mbytes_per_sec": 0, 00:16:04.266 "w_mbytes_per_sec": 0 00:16:04.266 }, 00:16:04.266 "claimed": true, 00:16:04.266 "claim_type": "exclusive_write", 00:16:04.266 "zoned": false, 00:16:04.266 "supported_io_types": { 00:16:04.266 "read": true, 00:16:04.266 "write": true, 00:16:04.266 "unmap": true, 00:16:04.266 "flush": true, 00:16:04.266 "reset": true, 00:16:04.266 "nvme_admin": false, 00:16:04.266 "nvme_io": false, 00:16:04.266 "nvme_io_md": false, 00:16:04.266 "write_zeroes": true, 00:16:04.266 "zcopy": true, 00:16:04.266 "get_zone_info": false, 00:16:04.266 "zone_management": false, 00:16:04.266 "zone_append": false, 00:16:04.266 "compare": false, 00:16:04.266 "compare_and_write": false, 00:16:04.266 "abort": true, 00:16:04.266 "seek_hole": false, 00:16:04.266 "seek_data": false, 00:16:04.266 "copy": true, 00:16:04.266 "nvme_iov_md": false 00:16:04.266 }, 00:16:04.266 "memory_domains": [ 00:16:04.266 { 00:16:04.266 "dma_device_id": "system", 00:16:04.266 "dma_device_type": 1 00:16:04.266 }, 00:16:04.266 { 00:16:04.266 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:04.266 "dma_device_type": 2 00:16:04.266 } 00:16:04.266 ], 00:16:04.266 "driver_specific": {} 00:16:04.266 } 00:16:04.266 ] 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:04.266 "name": "Existed_Raid", 00:16:04.266 "uuid": "5c8b56f3-99fb-4e43-af30-10fc26306a90", 00:16:04.266 "strip_size_kb": 0, 00:16:04.266 "state": "online", 00:16:04.266 "raid_level": "raid1", 00:16:04.266 "superblock": true, 00:16:04.266 "num_base_bdevs": 2, 00:16:04.266 "num_base_bdevs_discovered": 2, 00:16:04.266 "num_base_bdevs_operational": 2, 00:16:04.266 "base_bdevs_list": [ 00:16:04.266 { 00:16:04.266 "name": "BaseBdev1", 00:16:04.266 "uuid": "f92752e2-29da-49c0-a5b3-5922b6d45994", 00:16:04.266 "is_configured": true, 00:16:04.266 "data_offset": 256, 00:16:04.266 "data_size": 7936 00:16:04.266 }, 00:16:04.266 { 00:16:04.266 "name": "BaseBdev2", 00:16:04.266 "uuid": "d6ce9d07-2996-439c-ae06-bfe71ca55548", 00:16:04.266 "is_configured": true, 00:16:04.266 "data_offset": 256, 00:16:04.266 "data_size": 7936 00:16:04.266 } 00:16:04.266 ] 00:16:04.266 }' 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:04.266 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:04.526 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:04.526 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:04.526 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:04.526 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:04.526 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:04.526 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:04.526 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:04.526 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.526 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:04.526 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:04.526 [2024-12-07 05:43:37.883923] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:04.786 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.786 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:04.786 "name": "Existed_Raid", 00:16:04.786 "aliases": [ 00:16:04.786 "5c8b56f3-99fb-4e43-af30-10fc26306a90" 00:16:04.786 ], 00:16:04.786 "product_name": "Raid Volume", 00:16:04.786 "block_size": 4128, 00:16:04.786 "num_blocks": 7936, 00:16:04.786 "uuid": "5c8b56f3-99fb-4e43-af30-10fc26306a90", 00:16:04.786 "md_size": 32, 00:16:04.786 "md_interleave": true, 00:16:04.786 "dif_type": 0, 00:16:04.786 "assigned_rate_limits": { 00:16:04.786 "rw_ios_per_sec": 0, 00:16:04.786 "rw_mbytes_per_sec": 0, 00:16:04.786 "r_mbytes_per_sec": 0, 00:16:04.786 "w_mbytes_per_sec": 0 00:16:04.786 }, 00:16:04.786 "claimed": false, 00:16:04.786 "zoned": false, 00:16:04.786 "supported_io_types": { 00:16:04.786 "read": true, 00:16:04.786 "write": true, 00:16:04.786 "unmap": false, 00:16:04.786 "flush": false, 00:16:04.786 "reset": true, 00:16:04.786 "nvme_admin": false, 00:16:04.786 "nvme_io": false, 00:16:04.786 "nvme_io_md": false, 00:16:04.786 "write_zeroes": true, 00:16:04.786 "zcopy": false, 00:16:04.786 "get_zone_info": false, 00:16:04.786 "zone_management": false, 00:16:04.786 "zone_append": false, 00:16:04.786 "compare": false, 00:16:04.786 "compare_and_write": false, 00:16:04.786 "abort": false, 00:16:04.786 "seek_hole": false, 00:16:04.786 "seek_data": false, 00:16:04.786 "copy": false, 00:16:04.787 "nvme_iov_md": false 00:16:04.787 }, 00:16:04.787 "memory_domains": [ 00:16:04.787 { 00:16:04.787 "dma_device_id": "system", 00:16:04.787 "dma_device_type": 1 00:16:04.787 }, 00:16:04.787 { 00:16:04.787 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:04.787 "dma_device_type": 2 00:16:04.787 }, 00:16:04.787 { 00:16:04.787 "dma_device_id": "system", 00:16:04.787 "dma_device_type": 1 00:16:04.787 }, 00:16:04.787 { 00:16:04.787 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:04.787 "dma_device_type": 2 00:16:04.787 } 00:16:04.787 ], 00:16:04.787 "driver_specific": { 00:16:04.787 "raid": { 00:16:04.787 "uuid": "5c8b56f3-99fb-4e43-af30-10fc26306a90", 00:16:04.787 "strip_size_kb": 0, 00:16:04.787 "state": "online", 00:16:04.787 "raid_level": "raid1", 00:16:04.787 "superblock": true, 00:16:04.787 "num_base_bdevs": 2, 00:16:04.787 "num_base_bdevs_discovered": 2, 00:16:04.787 "num_base_bdevs_operational": 2, 00:16:04.787 "base_bdevs_list": [ 00:16:04.787 { 00:16:04.787 "name": "BaseBdev1", 00:16:04.787 "uuid": "f92752e2-29da-49c0-a5b3-5922b6d45994", 00:16:04.787 "is_configured": true, 00:16:04.787 "data_offset": 256, 00:16:04.787 "data_size": 7936 00:16:04.787 }, 00:16:04.787 { 00:16:04.787 "name": "BaseBdev2", 00:16:04.787 "uuid": "d6ce9d07-2996-439c-ae06-bfe71ca55548", 00:16:04.787 "is_configured": true, 00:16:04.787 "data_offset": 256, 00:16:04.787 "data_size": 7936 00:16:04.787 } 00:16:04.787 ] 00:16:04.787 } 00:16:04.787 } 00:16:04.787 }' 00:16:04.787 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:04.787 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:04.787 BaseBdev2' 00:16:04.787 05:43:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:04.787 [2024-12-07 05:43:38.083374] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:04.787 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.047 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:05.047 "name": "Existed_Raid", 00:16:05.047 "uuid": "5c8b56f3-99fb-4e43-af30-10fc26306a90", 00:16:05.047 "strip_size_kb": 0, 00:16:05.047 "state": "online", 00:16:05.047 "raid_level": "raid1", 00:16:05.047 "superblock": true, 00:16:05.047 "num_base_bdevs": 2, 00:16:05.047 "num_base_bdevs_discovered": 1, 00:16:05.047 "num_base_bdevs_operational": 1, 00:16:05.047 "base_bdevs_list": [ 00:16:05.047 { 00:16:05.047 "name": null, 00:16:05.047 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:05.047 "is_configured": false, 00:16:05.047 "data_offset": 0, 00:16:05.047 "data_size": 7936 00:16:05.047 }, 00:16:05.047 { 00:16:05.047 "name": "BaseBdev2", 00:16:05.047 "uuid": "d6ce9d07-2996-439c-ae06-bfe71ca55548", 00:16:05.047 "is_configured": true, 00:16:05.047 "data_offset": 256, 00:16:05.047 "data_size": 7936 00:16:05.047 } 00:16:05.047 ] 00:16:05.047 }' 00:16:05.047 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:05.047 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:05.308 [2024-12-07 05:43:38.518142] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:05.308 [2024-12-07 05:43:38.518247] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:05.308 [2024-12-07 05:43:38.530459] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:05.308 [2024-12-07 05:43:38.530503] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:05.308 [2024-12-07 05:43:38.530515] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 98341 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 98341 ']' 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 98341 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 98341 00:16:05.308 killing process with pid 98341 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 98341' 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 98341 00:16:05.308 [2024-12-07 05:43:38.604074] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:05.308 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 98341 00:16:05.308 [2024-12-07 05:43:38.605032] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:05.569 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:16:05.569 00:16:05.569 real 0m3.712s 00:16:05.569 user 0m5.843s 00:16:05.569 sys 0m0.760s 00:16:05.569 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:05.569 05:43:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:05.569 ************************************ 00:16:05.569 END TEST raid_state_function_test_sb_md_interleaved 00:16:05.569 ************************************ 00:16:05.569 05:43:38 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:16:05.569 05:43:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:16:05.569 05:43:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:05.569 05:43:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:05.569 ************************************ 00:16:05.569 START TEST raid_superblock_test_md_interleaved 00:16:05.569 ************************************ 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=98578 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 98578 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 98578 ']' 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:05.569 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:05.569 05:43:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:05.829 [2024-12-07 05:43:38.963805] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:16:05.829 [2024-12-07 05:43:38.963953] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid98578 ] 00:16:05.829 [2024-12-07 05:43:39.112755] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:05.829 [2024-12-07 05:43:39.137081] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:05.829 [2024-12-07 05:43:39.178979] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:05.829 [2024-12-07 05:43:39.179021] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.826 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:06.827 malloc1 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:06.827 [2024-12-07 05:43:39.849922] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:06.827 [2024-12-07 05:43:39.849985] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:06.827 [2024-12-07 05:43:39.850008] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:06.827 [2024-12-07 05:43:39.850025] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:06.827 [2024-12-07 05:43:39.851912] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:06.827 [2024-12-07 05:43:39.851948] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:06.827 pt1 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:06.827 malloc2 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:06.827 [2024-12-07 05:43:39.882528] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:06.827 [2024-12-07 05:43:39.882576] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:06.827 [2024-12-07 05:43:39.882592] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:06.827 [2024-12-07 05:43:39.882602] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:06.827 [2024-12-07 05:43:39.884414] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:06.827 [2024-12-07 05:43:39.884446] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:06.827 pt2 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:06.827 [2024-12-07 05:43:39.894535] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:06.827 [2024-12-07 05:43:39.896332] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:06.827 [2024-12-07 05:43:39.896497] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:06.827 [2024-12-07 05:43:39.896521] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:06.827 [2024-12-07 05:43:39.896604] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:06.827 [2024-12-07 05:43:39.896694] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:06.827 [2024-12-07 05:43:39.896705] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:06.827 [2024-12-07 05:43:39.896767] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.827 "name": "raid_bdev1", 00:16:06.827 "uuid": "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad", 00:16:06.827 "strip_size_kb": 0, 00:16:06.827 "state": "online", 00:16:06.827 "raid_level": "raid1", 00:16:06.827 "superblock": true, 00:16:06.827 "num_base_bdevs": 2, 00:16:06.827 "num_base_bdevs_discovered": 2, 00:16:06.827 "num_base_bdevs_operational": 2, 00:16:06.827 "base_bdevs_list": [ 00:16:06.827 { 00:16:06.827 "name": "pt1", 00:16:06.827 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:06.827 "is_configured": true, 00:16:06.827 "data_offset": 256, 00:16:06.827 "data_size": 7936 00:16:06.827 }, 00:16:06.827 { 00:16:06.827 "name": "pt2", 00:16:06.827 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:06.827 "is_configured": true, 00:16:06.827 "data_offset": 256, 00:16:06.827 "data_size": 7936 00:16:06.827 } 00:16:06.827 ] 00:16:06.827 }' 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.827 05:43:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:07.099 [2024-12-07 05:43:40.338083] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.099 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:07.099 "name": "raid_bdev1", 00:16:07.099 "aliases": [ 00:16:07.099 "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad" 00:16:07.099 ], 00:16:07.099 "product_name": "Raid Volume", 00:16:07.099 "block_size": 4128, 00:16:07.099 "num_blocks": 7936, 00:16:07.099 "uuid": "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad", 00:16:07.099 "md_size": 32, 00:16:07.099 "md_interleave": true, 00:16:07.099 "dif_type": 0, 00:16:07.099 "assigned_rate_limits": { 00:16:07.099 "rw_ios_per_sec": 0, 00:16:07.099 "rw_mbytes_per_sec": 0, 00:16:07.099 "r_mbytes_per_sec": 0, 00:16:07.099 "w_mbytes_per_sec": 0 00:16:07.099 }, 00:16:07.099 "claimed": false, 00:16:07.099 "zoned": false, 00:16:07.099 "supported_io_types": { 00:16:07.099 "read": true, 00:16:07.099 "write": true, 00:16:07.099 "unmap": false, 00:16:07.099 "flush": false, 00:16:07.099 "reset": true, 00:16:07.099 "nvme_admin": false, 00:16:07.099 "nvme_io": false, 00:16:07.099 "nvme_io_md": false, 00:16:07.099 "write_zeroes": true, 00:16:07.099 "zcopy": false, 00:16:07.099 "get_zone_info": false, 00:16:07.099 "zone_management": false, 00:16:07.099 "zone_append": false, 00:16:07.099 "compare": false, 00:16:07.099 "compare_and_write": false, 00:16:07.099 "abort": false, 00:16:07.099 "seek_hole": false, 00:16:07.099 "seek_data": false, 00:16:07.099 "copy": false, 00:16:07.099 "nvme_iov_md": false 00:16:07.099 }, 00:16:07.099 "memory_domains": [ 00:16:07.099 { 00:16:07.099 "dma_device_id": "system", 00:16:07.099 "dma_device_type": 1 00:16:07.099 }, 00:16:07.099 { 00:16:07.099 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:07.099 "dma_device_type": 2 00:16:07.099 }, 00:16:07.099 { 00:16:07.099 "dma_device_id": "system", 00:16:07.099 "dma_device_type": 1 00:16:07.100 }, 00:16:07.100 { 00:16:07.100 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:07.100 "dma_device_type": 2 00:16:07.100 } 00:16:07.100 ], 00:16:07.100 "driver_specific": { 00:16:07.100 "raid": { 00:16:07.100 "uuid": "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad", 00:16:07.100 "strip_size_kb": 0, 00:16:07.100 "state": "online", 00:16:07.100 "raid_level": "raid1", 00:16:07.100 "superblock": true, 00:16:07.100 "num_base_bdevs": 2, 00:16:07.100 "num_base_bdevs_discovered": 2, 00:16:07.100 "num_base_bdevs_operational": 2, 00:16:07.100 "base_bdevs_list": [ 00:16:07.100 { 00:16:07.100 "name": "pt1", 00:16:07.100 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:07.100 "is_configured": true, 00:16:07.100 "data_offset": 256, 00:16:07.100 "data_size": 7936 00:16:07.100 }, 00:16:07.100 { 00:16:07.100 "name": "pt2", 00:16:07.100 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:07.100 "is_configured": true, 00:16:07.100 "data_offset": 256, 00:16:07.100 "data_size": 7936 00:16:07.100 } 00:16:07.100 ] 00:16:07.100 } 00:16:07.100 } 00:16:07.100 }' 00:16:07.100 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:07.100 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:07.100 pt2' 00:16:07.100 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.360 [2024-12-07 05:43:40.569569] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad ']' 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.360 [2024-12-07 05:43:40.601314] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:07.360 [2024-12-07 05:43:40.601341] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:07.360 [2024-12-07 05:43:40.601403] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:07.360 [2024-12-07 05:43:40.601466] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:07.360 [2024-12-07 05:43:40.601476] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.360 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:07.361 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.620 [2024-12-07 05:43:40.733090] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:07.620 [2024-12-07 05:43:40.734976] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:07.620 [2024-12-07 05:43:40.735044] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:07.620 [2024-12-07 05:43:40.735086] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:07.620 [2024-12-07 05:43:40.735104] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:07.620 [2024-12-07 05:43:40.735113] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:16:07.620 request: 00:16:07.620 { 00:16:07.620 "name": "raid_bdev1", 00:16:07.620 "raid_level": "raid1", 00:16:07.620 "base_bdevs": [ 00:16:07.620 "malloc1", 00:16:07.620 "malloc2" 00:16:07.620 ], 00:16:07.620 "superblock": false, 00:16:07.620 "method": "bdev_raid_create", 00:16:07.620 "req_id": 1 00:16:07.620 } 00:16:07.620 Got JSON-RPC error response 00:16:07.620 response: 00:16:07.620 { 00:16:07.620 "code": -17, 00:16:07.620 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:07.620 } 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:07.620 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.621 [2024-12-07 05:43:40.796962] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:07.621 [2024-12-07 05:43:40.797011] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:07.621 [2024-12-07 05:43:40.797028] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:07.621 [2024-12-07 05:43:40.797036] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:07.621 [2024-12-07 05:43:40.798917] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:07.621 [2024-12-07 05:43:40.798948] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:07.621 [2024-12-07 05:43:40.798994] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:07.621 [2024-12-07 05:43:40.799032] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:07.621 pt1 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:07.621 "name": "raid_bdev1", 00:16:07.621 "uuid": "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad", 00:16:07.621 "strip_size_kb": 0, 00:16:07.621 "state": "configuring", 00:16:07.621 "raid_level": "raid1", 00:16:07.621 "superblock": true, 00:16:07.621 "num_base_bdevs": 2, 00:16:07.621 "num_base_bdevs_discovered": 1, 00:16:07.621 "num_base_bdevs_operational": 2, 00:16:07.621 "base_bdevs_list": [ 00:16:07.621 { 00:16:07.621 "name": "pt1", 00:16:07.621 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:07.621 "is_configured": true, 00:16:07.621 "data_offset": 256, 00:16:07.621 "data_size": 7936 00:16:07.621 }, 00:16:07.621 { 00:16:07.621 "name": null, 00:16:07.621 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:07.621 "is_configured": false, 00:16:07.621 "data_offset": 256, 00:16:07.621 "data_size": 7936 00:16:07.621 } 00:16:07.621 ] 00:16:07.621 }' 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:07.621 05:43:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:08.189 [2024-12-07 05:43:41.272147] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:08.189 [2024-12-07 05:43:41.272193] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:08.189 [2024-12-07 05:43:41.272213] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:08.189 [2024-12-07 05:43:41.272221] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:08.189 [2024-12-07 05:43:41.272391] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:08.189 [2024-12-07 05:43:41.272405] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:08.189 [2024-12-07 05:43:41.272450] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:08.189 [2024-12-07 05:43:41.272470] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:08.189 [2024-12-07 05:43:41.272544] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:08.189 [2024-12-07 05:43:41.272551] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:08.189 [2024-12-07 05:43:41.272619] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:08.189 [2024-12-07 05:43:41.272705] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:08.189 [2024-12-07 05:43:41.272717] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:16:08.189 [2024-12-07 05:43:41.272771] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:08.189 pt2 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:08.189 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:08.190 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:08.190 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:08.190 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.190 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.190 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.190 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:08.190 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.190 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:08.190 "name": "raid_bdev1", 00:16:08.190 "uuid": "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad", 00:16:08.190 "strip_size_kb": 0, 00:16:08.190 "state": "online", 00:16:08.190 "raid_level": "raid1", 00:16:08.190 "superblock": true, 00:16:08.190 "num_base_bdevs": 2, 00:16:08.190 "num_base_bdevs_discovered": 2, 00:16:08.190 "num_base_bdevs_operational": 2, 00:16:08.190 "base_bdevs_list": [ 00:16:08.190 { 00:16:08.190 "name": "pt1", 00:16:08.190 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:08.190 "is_configured": true, 00:16:08.190 "data_offset": 256, 00:16:08.190 "data_size": 7936 00:16:08.190 }, 00:16:08.190 { 00:16:08.190 "name": "pt2", 00:16:08.190 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:08.190 "is_configured": true, 00:16:08.190 "data_offset": 256, 00:16:08.190 "data_size": 7936 00:16:08.190 } 00:16:08.190 ] 00:16:08.190 }' 00:16:08.190 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:08.190 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:08.450 [2024-12-07 05:43:41.691705] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:08.450 "name": "raid_bdev1", 00:16:08.450 "aliases": [ 00:16:08.450 "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad" 00:16:08.450 ], 00:16:08.450 "product_name": "Raid Volume", 00:16:08.450 "block_size": 4128, 00:16:08.450 "num_blocks": 7936, 00:16:08.450 "uuid": "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad", 00:16:08.450 "md_size": 32, 00:16:08.450 "md_interleave": true, 00:16:08.450 "dif_type": 0, 00:16:08.450 "assigned_rate_limits": { 00:16:08.450 "rw_ios_per_sec": 0, 00:16:08.450 "rw_mbytes_per_sec": 0, 00:16:08.450 "r_mbytes_per_sec": 0, 00:16:08.450 "w_mbytes_per_sec": 0 00:16:08.450 }, 00:16:08.450 "claimed": false, 00:16:08.450 "zoned": false, 00:16:08.450 "supported_io_types": { 00:16:08.450 "read": true, 00:16:08.450 "write": true, 00:16:08.450 "unmap": false, 00:16:08.450 "flush": false, 00:16:08.450 "reset": true, 00:16:08.450 "nvme_admin": false, 00:16:08.450 "nvme_io": false, 00:16:08.450 "nvme_io_md": false, 00:16:08.450 "write_zeroes": true, 00:16:08.450 "zcopy": false, 00:16:08.450 "get_zone_info": false, 00:16:08.450 "zone_management": false, 00:16:08.450 "zone_append": false, 00:16:08.450 "compare": false, 00:16:08.450 "compare_and_write": false, 00:16:08.450 "abort": false, 00:16:08.450 "seek_hole": false, 00:16:08.450 "seek_data": false, 00:16:08.450 "copy": false, 00:16:08.450 "nvme_iov_md": false 00:16:08.450 }, 00:16:08.450 "memory_domains": [ 00:16:08.450 { 00:16:08.450 "dma_device_id": "system", 00:16:08.450 "dma_device_type": 1 00:16:08.450 }, 00:16:08.450 { 00:16:08.450 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:08.450 "dma_device_type": 2 00:16:08.450 }, 00:16:08.450 { 00:16:08.450 "dma_device_id": "system", 00:16:08.450 "dma_device_type": 1 00:16:08.450 }, 00:16:08.450 { 00:16:08.450 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:08.450 "dma_device_type": 2 00:16:08.450 } 00:16:08.450 ], 00:16:08.450 "driver_specific": { 00:16:08.450 "raid": { 00:16:08.450 "uuid": "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad", 00:16:08.450 "strip_size_kb": 0, 00:16:08.450 "state": "online", 00:16:08.450 "raid_level": "raid1", 00:16:08.450 "superblock": true, 00:16:08.450 "num_base_bdevs": 2, 00:16:08.450 "num_base_bdevs_discovered": 2, 00:16:08.450 "num_base_bdevs_operational": 2, 00:16:08.450 "base_bdevs_list": [ 00:16:08.450 { 00:16:08.450 "name": "pt1", 00:16:08.450 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:08.450 "is_configured": true, 00:16:08.450 "data_offset": 256, 00:16:08.450 "data_size": 7936 00:16:08.450 }, 00:16:08.450 { 00:16:08.450 "name": "pt2", 00:16:08.450 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:08.450 "is_configured": true, 00:16:08.450 "data_offset": 256, 00:16:08.450 "data_size": 7936 00:16:08.450 } 00:16:08.450 ] 00:16:08.450 } 00:16:08.450 } 00:16:08.450 }' 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:08.450 pt2' 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.450 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:08.711 [2024-12-07 05:43:41.911299] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad '!=' fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad ']' 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:08.711 [2024-12-07 05:43:41.947034] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:08.711 05:43:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.711 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:08.711 "name": "raid_bdev1", 00:16:08.711 "uuid": "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad", 00:16:08.711 "strip_size_kb": 0, 00:16:08.711 "state": "online", 00:16:08.711 "raid_level": "raid1", 00:16:08.711 "superblock": true, 00:16:08.711 "num_base_bdevs": 2, 00:16:08.711 "num_base_bdevs_discovered": 1, 00:16:08.711 "num_base_bdevs_operational": 1, 00:16:08.711 "base_bdevs_list": [ 00:16:08.711 { 00:16:08.711 "name": null, 00:16:08.711 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:08.711 "is_configured": false, 00:16:08.711 "data_offset": 0, 00:16:08.711 "data_size": 7936 00:16:08.711 }, 00:16:08.711 { 00:16:08.711 "name": "pt2", 00:16:08.711 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:08.712 "is_configured": true, 00:16:08.712 "data_offset": 256, 00:16:08.712 "data_size": 7936 00:16:08.712 } 00:16:08.712 ] 00:16:08.712 }' 00:16:08.712 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:08.712 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:09.279 [2024-12-07 05:43:42.406234] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:09.279 [2024-12-07 05:43:42.406263] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:09.279 [2024-12-07 05:43:42.406325] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:09.279 [2024-12-07 05:43:42.406375] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:09.279 [2024-12-07 05:43:42.406384] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.279 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:09.279 [2024-12-07 05:43:42.478098] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:09.279 [2024-12-07 05:43:42.478139] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:09.279 [2024-12-07 05:43:42.478156] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:16:09.279 [2024-12-07 05:43:42.478164] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:09.279 [2024-12-07 05:43:42.480026] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:09.279 [2024-12-07 05:43:42.480058] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:09.279 [2024-12-07 05:43:42.480106] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:09.279 [2024-12-07 05:43:42.480146] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:09.279 [2024-12-07 05:43:42.480211] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:16:09.279 [2024-12-07 05:43:42.480223] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:09.279 [2024-12-07 05:43:42.480287] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:16:09.280 [2024-12-07 05:43:42.480338] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:16:09.280 [2024-12-07 05:43:42.480347] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:16:09.280 [2024-12-07 05:43:42.480397] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:09.280 pt2 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:09.280 "name": "raid_bdev1", 00:16:09.280 "uuid": "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad", 00:16:09.280 "strip_size_kb": 0, 00:16:09.280 "state": "online", 00:16:09.280 "raid_level": "raid1", 00:16:09.280 "superblock": true, 00:16:09.280 "num_base_bdevs": 2, 00:16:09.280 "num_base_bdevs_discovered": 1, 00:16:09.280 "num_base_bdevs_operational": 1, 00:16:09.280 "base_bdevs_list": [ 00:16:09.280 { 00:16:09.280 "name": null, 00:16:09.280 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:09.280 "is_configured": false, 00:16:09.280 "data_offset": 256, 00:16:09.280 "data_size": 7936 00:16:09.280 }, 00:16:09.280 { 00:16:09.280 "name": "pt2", 00:16:09.280 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:09.280 "is_configured": true, 00:16:09.280 "data_offset": 256, 00:16:09.280 "data_size": 7936 00:16:09.280 } 00:16:09.280 ] 00:16:09.280 }' 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:09.280 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:09.538 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:09.538 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.538 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:09.538 [2024-12-07 05:43:42.873454] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:09.538 [2024-12-07 05:43:42.873482] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:09.538 [2024-12-07 05:43:42.873546] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:09.538 [2024-12-07 05:43:42.873589] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:09.538 [2024-12-07 05:43:42.873602] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:16:09.538 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.538 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.538 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:09.538 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.538 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:09.538 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:09.796 [2024-12-07 05:43:42.933369] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:09.796 [2024-12-07 05:43:42.933432] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:09.796 [2024-12-07 05:43:42.933447] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:16:09.796 [2024-12-07 05:43:42.933459] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:09.796 [2024-12-07 05:43:42.935407] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:09.796 [2024-12-07 05:43:42.935441] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:09.796 [2024-12-07 05:43:42.935490] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:09.796 [2024-12-07 05:43:42.935524] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:09.796 [2024-12-07 05:43:42.935596] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:09.796 [2024-12-07 05:43:42.935620] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:09.796 [2024-12-07 05:43:42.935651] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:16:09.796 [2024-12-07 05:43:42.935689] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:09.796 [2024-12-07 05:43:42.935752] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:16:09.796 [2024-12-07 05:43:42.935763] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:09.796 [2024-12-07 05:43:42.935846] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:09.796 [2024-12-07 05:43:42.935899] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:16:09.796 [2024-12-07 05:43:42.935912] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:16:09.796 [2024-12-07 05:43:42.935974] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:09.796 pt1 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.796 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:09.797 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.797 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:09.797 "name": "raid_bdev1", 00:16:09.797 "uuid": "fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad", 00:16:09.797 "strip_size_kb": 0, 00:16:09.797 "state": "online", 00:16:09.797 "raid_level": "raid1", 00:16:09.797 "superblock": true, 00:16:09.797 "num_base_bdevs": 2, 00:16:09.797 "num_base_bdevs_discovered": 1, 00:16:09.797 "num_base_bdevs_operational": 1, 00:16:09.797 "base_bdevs_list": [ 00:16:09.797 { 00:16:09.797 "name": null, 00:16:09.797 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:09.797 "is_configured": false, 00:16:09.797 "data_offset": 256, 00:16:09.797 "data_size": 7936 00:16:09.797 }, 00:16:09.797 { 00:16:09.797 "name": "pt2", 00:16:09.797 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:09.797 "is_configured": true, 00:16:09.797 "data_offset": 256, 00:16:09.797 "data_size": 7936 00:16:09.797 } 00:16:09.797 ] 00:16:09.797 }' 00:16:09.797 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:09.797 05:43:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:10.055 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:10.055 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:10.055 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.055 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:10.055 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:10.314 [2024-12-07 05:43:43.432771] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad '!=' fcaa1ca7-1efc-44eb-a2c7-cf2631b79fad ']' 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 98578 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 98578 ']' 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 98578 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 98578 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:10.314 killing process with pid 98578 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 98578' 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@973 -- # kill 98578 00:16:10.314 [2024-12-07 05:43:43.503502] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:10.314 [2024-12-07 05:43:43.503569] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:10.314 [2024-12-07 05:43:43.503616] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:10.314 [2024-12-07 05:43:43.503639] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:16:10.314 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@978 -- # wait 98578 00:16:10.314 [2024-12-07 05:43:43.527062] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:10.572 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:16:10.572 00:16:10.572 real 0m4.860s 00:16:10.572 user 0m8.027s 00:16:10.572 sys 0m0.933s 00:16:10.572 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:10.572 05:43:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:10.572 ************************************ 00:16:10.572 END TEST raid_superblock_test_md_interleaved 00:16:10.572 ************************************ 00:16:10.572 05:43:43 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:16:10.572 05:43:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:10.572 05:43:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:10.572 05:43:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:10.572 ************************************ 00:16:10.572 START TEST raid_rebuild_test_sb_md_interleaved 00:16:10.572 ************************************ 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false false 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=98890 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 98890 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 98890 ']' 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:10.572 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:10.572 05:43:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:10.572 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:10.572 Zero copy mechanism will not be used. 00:16:10.572 [2024-12-07 05:43:43.899539] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:16:10.573 [2024-12-07 05:43:43.899719] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid98890 ] 00:16:10.832 [2024-12-07 05:43:44.039823] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:10.832 [2024-12-07 05:43:44.064626] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:10.832 [2024-12-07 05:43:44.106770] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:10.832 [2024-12-07 05:43:44.106806] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:11.402 BaseBdev1_malloc 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:11.402 [2024-12-07 05:43:44.761784] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:11.402 [2024-12-07 05:43:44.761839] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:11.402 [2024-12-07 05:43:44.761866] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:11.402 [2024-12-07 05:43:44.761875] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:11.402 [2024-12-07 05:43:44.763730] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:11.402 [2024-12-07 05:43:44.763762] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:11.402 BaseBdev1 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.402 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:11.663 BaseBdev2_malloc 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:11.663 [2024-12-07 05:43:44.790281] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:11.663 [2024-12-07 05:43:44.790327] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:11.663 [2024-12-07 05:43:44.790346] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:11.663 [2024-12-07 05:43:44.790354] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:11.663 [2024-12-07 05:43:44.792148] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:11.663 [2024-12-07 05:43:44.792185] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:11.663 BaseBdev2 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:11.663 spare_malloc 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:11.663 spare_delay 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:11.663 [2024-12-07 05:43:44.840321] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:11.663 [2024-12-07 05:43:44.840367] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:11.663 [2024-12-07 05:43:44.840386] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:11.663 [2024-12-07 05:43:44.840394] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:11.663 [2024-12-07 05:43:44.842190] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:11.663 [2024-12-07 05:43:44.842222] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:11.663 spare 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:11.663 [2024-12-07 05:43:44.852333] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:11.663 [2024-12-07 05:43:44.854122] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:11.663 [2024-12-07 05:43:44.854285] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:11.663 [2024-12-07 05:43:44.854306] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:11.663 [2024-12-07 05:43:44.854411] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:11.663 [2024-12-07 05:43:44.854492] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:11.663 [2024-12-07 05:43:44.854504] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:11.663 [2024-12-07 05:43:44.854571] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.663 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:11.663 "name": "raid_bdev1", 00:16:11.663 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:11.663 "strip_size_kb": 0, 00:16:11.663 "state": "online", 00:16:11.663 "raid_level": "raid1", 00:16:11.663 "superblock": true, 00:16:11.663 "num_base_bdevs": 2, 00:16:11.663 "num_base_bdevs_discovered": 2, 00:16:11.663 "num_base_bdevs_operational": 2, 00:16:11.663 "base_bdevs_list": [ 00:16:11.663 { 00:16:11.663 "name": "BaseBdev1", 00:16:11.664 "uuid": "d1a0b78c-fd75-50f6-bb88-72f71dcb9516", 00:16:11.664 "is_configured": true, 00:16:11.664 "data_offset": 256, 00:16:11.664 "data_size": 7936 00:16:11.664 }, 00:16:11.664 { 00:16:11.664 "name": "BaseBdev2", 00:16:11.664 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:11.664 "is_configured": true, 00:16:11.664 "data_offset": 256, 00:16:11.664 "data_size": 7936 00:16:11.664 } 00:16:11.664 ] 00:16:11.664 }' 00:16:11.664 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:11.664 05:43:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:12.232 [2024-12-07 05:43:45.307826] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:12.232 [2024-12-07 05:43:45.379399] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:12.232 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.233 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:12.233 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.233 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:12.233 "name": "raid_bdev1", 00:16:12.233 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:12.233 "strip_size_kb": 0, 00:16:12.233 "state": "online", 00:16:12.233 "raid_level": "raid1", 00:16:12.233 "superblock": true, 00:16:12.233 "num_base_bdevs": 2, 00:16:12.233 "num_base_bdevs_discovered": 1, 00:16:12.233 "num_base_bdevs_operational": 1, 00:16:12.233 "base_bdevs_list": [ 00:16:12.233 { 00:16:12.233 "name": null, 00:16:12.233 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:12.233 "is_configured": false, 00:16:12.233 "data_offset": 0, 00:16:12.233 "data_size": 7936 00:16:12.233 }, 00:16:12.233 { 00:16:12.233 "name": "BaseBdev2", 00:16:12.233 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:12.233 "is_configured": true, 00:16:12.233 "data_offset": 256, 00:16:12.233 "data_size": 7936 00:16:12.233 } 00:16:12.233 ] 00:16:12.233 }' 00:16:12.233 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:12.233 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:12.492 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:12.492 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.492 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:12.492 [2024-12-07 05:43:45.766758] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:12.492 [2024-12-07 05:43:45.770348] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:16:12.492 [2024-12-07 05:43:45.772228] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:12.492 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.492 05:43:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:13.430 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:13.430 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:13.430 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:13.430 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:13.430 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:13.430 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.430 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.430 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:13.431 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.690 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.690 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:13.690 "name": "raid_bdev1", 00:16:13.690 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:13.690 "strip_size_kb": 0, 00:16:13.690 "state": "online", 00:16:13.690 "raid_level": "raid1", 00:16:13.690 "superblock": true, 00:16:13.690 "num_base_bdevs": 2, 00:16:13.690 "num_base_bdevs_discovered": 2, 00:16:13.690 "num_base_bdevs_operational": 2, 00:16:13.690 "process": { 00:16:13.690 "type": "rebuild", 00:16:13.690 "target": "spare", 00:16:13.690 "progress": { 00:16:13.690 "blocks": 2560, 00:16:13.690 "percent": 32 00:16:13.690 } 00:16:13.690 }, 00:16:13.690 "base_bdevs_list": [ 00:16:13.690 { 00:16:13.690 "name": "spare", 00:16:13.690 "uuid": "b6f4fc60-d07d-58b9-9a25-1b58d45d7306", 00:16:13.690 "is_configured": true, 00:16:13.690 "data_offset": 256, 00:16:13.691 "data_size": 7936 00:16:13.691 }, 00:16:13.691 { 00:16:13.691 "name": "BaseBdev2", 00:16:13.691 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:13.691 "is_configured": true, 00:16:13.691 "data_offset": 256, 00:16:13.691 "data_size": 7936 00:16:13.691 } 00:16:13.691 ] 00:16:13.691 }' 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.691 [2024-12-07 05:43:46.938974] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:13.691 [2024-12-07 05:43:46.976947] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:13.691 [2024-12-07 05:43:46.977020] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:13.691 [2024-12-07 05:43:46.977037] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:13.691 [2024-12-07 05:43:46.977044] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:13.691 05:43:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.691 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.691 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:13.691 "name": "raid_bdev1", 00:16:13.691 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:13.691 "strip_size_kb": 0, 00:16:13.691 "state": "online", 00:16:13.691 "raid_level": "raid1", 00:16:13.691 "superblock": true, 00:16:13.691 "num_base_bdevs": 2, 00:16:13.691 "num_base_bdevs_discovered": 1, 00:16:13.691 "num_base_bdevs_operational": 1, 00:16:13.691 "base_bdevs_list": [ 00:16:13.691 { 00:16:13.691 "name": null, 00:16:13.691 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:13.691 "is_configured": false, 00:16:13.691 "data_offset": 0, 00:16:13.691 "data_size": 7936 00:16:13.691 }, 00:16:13.691 { 00:16:13.691 "name": "BaseBdev2", 00:16:13.691 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:13.691 "is_configured": true, 00:16:13.691 "data_offset": 256, 00:16:13.691 "data_size": 7936 00:16:13.691 } 00:16:13.691 ] 00:16:13.691 }' 00:16:13.691 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:13.691 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:14.262 "name": "raid_bdev1", 00:16:14.262 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:14.262 "strip_size_kb": 0, 00:16:14.262 "state": "online", 00:16:14.262 "raid_level": "raid1", 00:16:14.262 "superblock": true, 00:16:14.262 "num_base_bdevs": 2, 00:16:14.262 "num_base_bdevs_discovered": 1, 00:16:14.262 "num_base_bdevs_operational": 1, 00:16:14.262 "base_bdevs_list": [ 00:16:14.262 { 00:16:14.262 "name": null, 00:16:14.262 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:14.262 "is_configured": false, 00:16:14.262 "data_offset": 0, 00:16:14.262 "data_size": 7936 00:16:14.262 }, 00:16:14.262 { 00:16:14.262 "name": "BaseBdev2", 00:16:14.262 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:14.262 "is_configured": true, 00:16:14.262 "data_offset": 256, 00:16:14.262 "data_size": 7936 00:16:14.262 } 00:16:14.262 ] 00:16:14.262 }' 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.262 [2024-12-07 05:43:47.472225] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:14.262 [2024-12-07 05:43:47.475898] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:14.262 [2024-12-07 05:43:47.477774] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:14.262 05:43:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:15.203 "name": "raid_bdev1", 00:16:15.203 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:15.203 "strip_size_kb": 0, 00:16:15.203 "state": "online", 00:16:15.203 "raid_level": "raid1", 00:16:15.203 "superblock": true, 00:16:15.203 "num_base_bdevs": 2, 00:16:15.203 "num_base_bdevs_discovered": 2, 00:16:15.203 "num_base_bdevs_operational": 2, 00:16:15.203 "process": { 00:16:15.203 "type": "rebuild", 00:16:15.203 "target": "spare", 00:16:15.203 "progress": { 00:16:15.203 "blocks": 2560, 00:16:15.203 "percent": 32 00:16:15.203 } 00:16:15.203 }, 00:16:15.203 "base_bdevs_list": [ 00:16:15.203 { 00:16:15.203 "name": "spare", 00:16:15.203 "uuid": "b6f4fc60-d07d-58b9-9a25-1b58d45d7306", 00:16:15.203 "is_configured": true, 00:16:15.203 "data_offset": 256, 00:16:15.203 "data_size": 7936 00:16:15.203 }, 00:16:15.203 { 00:16:15.203 "name": "BaseBdev2", 00:16:15.203 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:15.203 "is_configured": true, 00:16:15.203 "data_offset": 256, 00:16:15.203 "data_size": 7936 00:16:15.203 } 00:16:15.203 ] 00:16:15.203 }' 00:16:15.203 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:15.463 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=605 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:15.463 "name": "raid_bdev1", 00:16:15.463 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:15.463 "strip_size_kb": 0, 00:16:15.463 "state": "online", 00:16:15.463 "raid_level": "raid1", 00:16:15.463 "superblock": true, 00:16:15.463 "num_base_bdevs": 2, 00:16:15.463 "num_base_bdevs_discovered": 2, 00:16:15.463 "num_base_bdevs_operational": 2, 00:16:15.463 "process": { 00:16:15.463 "type": "rebuild", 00:16:15.463 "target": "spare", 00:16:15.463 "progress": { 00:16:15.463 "blocks": 2816, 00:16:15.463 "percent": 35 00:16:15.463 } 00:16:15.463 }, 00:16:15.463 "base_bdevs_list": [ 00:16:15.463 { 00:16:15.463 "name": "spare", 00:16:15.463 "uuid": "b6f4fc60-d07d-58b9-9a25-1b58d45d7306", 00:16:15.463 "is_configured": true, 00:16:15.463 "data_offset": 256, 00:16:15.463 "data_size": 7936 00:16:15.463 }, 00:16:15.463 { 00:16:15.463 "name": "BaseBdev2", 00:16:15.463 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:15.463 "is_configured": true, 00:16:15.463 "data_offset": 256, 00:16:15.463 "data_size": 7936 00:16:15.463 } 00:16:15.463 ] 00:16:15.463 }' 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:15.463 05:43:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.845 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:16.845 "name": "raid_bdev1", 00:16:16.845 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:16.845 "strip_size_kb": 0, 00:16:16.845 "state": "online", 00:16:16.845 "raid_level": "raid1", 00:16:16.846 "superblock": true, 00:16:16.846 "num_base_bdevs": 2, 00:16:16.846 "num_base_bdevs_discovered": 2, 00:16:16.846 "num_base_bdevs_operational": 2, 00:16:16.846 "process": { 00:16:16.846 "type": "rebuild", 00:16:16.846 "target": "spare", 00:16:16.846 "progress": { 00:16:16.846 "blocks": 5888, 00:16:16.846 "percent": 74 00:16:16.846 } 00:16:16.846 }, 00:16:16.846 "base_bdevs_list": [ 00:16:16.846 { 00:16:16.846 "name": "spare", 00:16:16.846 "uuid": "b6f4fc60-d07d-58b9-9a25-1b58d45d7306", 00:16:16.846 "is_configured": true, 00:16:16.846 "data_offset": 256, 00:16:16.846 "data_size": 7936 00:16:16.846 }, 00:16:16.846 { 00:16:16.846 "name": "BaseBdev2", 00:16:16.846 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:16.846 "is_configured": true, 00:16:16.846 "data_offset": 256, 00:16:16.846 "data_size": 7936 00:16:16.846 } 00:16:16.846 ] 00:16:16.846 }' 00:16:16.846 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:16.846 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:16.846 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:16.846 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:16.846 05:43:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:17.414 [2024-12-07 05:43:50.588566] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:17.414 [2024-12-07 05:43:50.588732] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:17.414 [2024-12-07 05:43:50.588869] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:17.675 "name": "raid_bdev1", 00:16:17.675 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:17.675 "strip_size_kb": 0, 00:16:17.675 "state": "online", 00:16:17.675 "raid_level": "raid1", 00:16:17.675 "superblock": true, 00:16:17.675 "num_base_bdevs": 2, 00:16:17.675 "num_base_bdevs_discovered": 2, 00:16:17.675 "num_base_bdevs_operational": 2, 00:16:17.675 "base_bdevs_list": [ 00:16:17.675 { 00:16:17.675 "name": "spare", 00:16:17.675 "uuid": "b6f4fc60-d07d-58b9-9a25-1b58d45d7306", 00:16:17.675 "is_configured": true, 00:16:17.675 "data_offset": 256, 00:16:17.675 "data_size": 7936 00:16:17.675 }, 00:16:17.675 { 00:16:17.675 "name": "BaseBdev2", 00:16:17.675 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:17.675 "is_configured": true, 00:16:17.675 "data_offset": 256, 00:16:17.675 "data_size": 7936 00:16:17.675 } 00:16:17.675 ] 00:16:17.675 }' 00:16:17.675 05:43:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:17.675 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:17.675 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:17.675 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:17.675 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:16:17.675 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:17.675 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:17.675 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:17.675 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:17.675 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:17.934 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:17.934 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:17.935 "name": "raid_bdev1", 00:16:17.935 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:17.935 "strip_size_kb": 0, 00:16:17.935 "state": "online", 00:16:17.935 "raid_level": "raid1", 00:16:17.935 "superblock": true, 00:16:17.935 "num_base_bdevs": 2, 00:16:17.935 "num_base_bdevs_discovered": 2, 00:16:17.935 "num_base_bdevs_operational": 2, 00:16:17.935 "base_bdevs_list": [ 00:16:17.935 { 00:16:17.935 "name": "spare", 00:16:17.935 "uuid": "b6f4fc60-d07d-58b9-9a25-1b58d45d7306", 00:16:17.935 "is_configured": true, 00:16:17.935 "data_offset": 256, 00:16:17.935 "data_size": 7936 00:16:17.935 }, 00:16:17.935 { 00:16:17.935 "name": "BaseBdev2", 00:16:17.935 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:17.935 "is_configured": true, 00:16:17.935 "data_offset": 256, 00:16:17.935 "data_size": 7936 00:16:17.935 } 00:16:17.935 ] 00:16:17.935 }' 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:17.935 "name": "raid_bdev1", 00:16:17.935 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:17.935 "strip_size_kb": 0, 00:16:17.935 "state": "online", 00:16:17.935 "raid_level": "raid1", 00:16:17.935 "superblock": true, 00:16:17.935 "num_base_bdevs": 2, 00:16:17.935 "num_base_bdevs_discovered": 2, 00:16:17.935 "num_base_bdevs_operational": 2, 00:16:17.935 "base_bdevs_list": [ 00:16:17.935 { 00:16:17.935 "name": "spare", 00:16:17.935 "uuid": "b6f4fc60-d07d-58b9-9a25-1b58d45d7306", 00:16:17.935 "is_configured": true, 00:16:17.935 "data_offset": 256, 00:16:17.935 "data_size": 7936 00:16:17.935 }, 00:16:17.935 { 00:16:17.935 "name": "BaseBdev2", 00:16:17.935 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:17.935 "is_configured": true, 00:16:17.935 "data_offset": 256, 00:16:17.935 "data_size": 7936 00:16:17.935 } 00:16:17.935 ] 00:16:17.935 }' 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:17.935 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.504 [2024-12-07 05:43:51.594706] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:18.504 [2024-12-07 05:43:51.594733] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:18.504 [2024-12-07 05:43:51.594826] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:18.504 [2024-12-07 05:43:51.594898] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:18.504 [2024-12-07 05:43:51.594909] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.504 [2024-12-07 05:43:51.646717] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:18.504 [2024-12-07 05:43:51.646813] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:18.504 [2024-12-07 05:43:51.646851] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:16:18.504 [2024-12-07 05:43:51.646864] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:18.504 [2024-12-07 05:43:51.648753] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:18.504 [2024-12-07 05:43:51.648792] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:18.504 [2024-12-07 05:43:51.648844] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:18.504 [2024-12-07 05:43:51.648883] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:18.504 [2024-12-07 05:43:51.648963] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:18.504 spare 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.504 [2024-12-07 05:43:51.748855] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:16:18.504 [2024-12-07 05:43:51.748915] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:18.504 [2024-12-07 05:43:51.749042] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:16:18.504 [2024-12-07 05:43:51.749169] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:16:18.504 [2024-12-07 05:43:51.749209] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:16:18.504 [2024-12-07 05:43:51.749315] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:18.504 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.505 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.505 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.505 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:18.505 "name": "raid_bdev1", 00:16:18.505 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:18.505 "strip_size_kb": 0, 00:16:18.505 "state": "online", 00:16:18.505 "raid_level": "raid1", 00:16:18.505 "superblock": true, 00:16:18.505 "num_base_bdevs": 2, 00:16:18.505 "num_base_bdevs_discovered": 2, 00:16:18.505 "num_base_bdevs_operational": 2, 00:16:18.505 "base_bdevs_list": [ 00:16:18.505 { 00:16:18.505 "name": "spare", 00:16:18.505 "uuid": "b6f4fc60-d07d-58b9-9a25-1b58d45d7306", 00:16:18.505 "is_configured": true, 00:16:18.505 "data_offset": 256, 00:16:18.505 "data_size": 7936 00:16:18.505 }, 00:16:18.505 { 00:16:18.505 "name": "BaseBdev2", 00:16:18.505 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:18.505 "is_configured": true, 00:16:18.505 "data_offset": 256, 00:16:18.505 "data_size": 7936 00:16:18.505 } 00:16:18.505 ] 00:16:18.505 }' 00:16:18.505 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:18.505 05:43:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:19.075 "name": "raid_bdev1", 00:16:19.075 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:19.075 "strip_size_kb": 0, 00:16:19.075 "state": "online", 00:16:19.075 "raid_level": "raid1", 00:16:19.075 "superblock": true, 00:16:19.075 "num_base_bdevs": 2, 00:16:19.075 "num_base_bdevs_discovered": 2, 00:16:19.075 "num_base_bdevs_operational": 2, 00:16:19.075 "base_bdevs_list": [ 00:16:19.075 { 00:16:19.075 "name": "spare", 00:16:19.075 "uuid": "b6f4fc60-d07d-58b9-9a25-1b58d45d7306", 00:16:19.075 "is_configured": true, 00:16:19.075 "data_offset": 256, 00:16:19.075 "data_size": 7936 00:16:19.075 }, 00:16:19.075 { 00:16:19.075 "name": "BaseBdev2", 00:16:19.075 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:19.075 "is_configured": true, 00:16:19.075 "data_offset": 256, 00:16:19.075 "data_size": 7936 00:16:19.075 } 00:16:19.075 ] 00:16:19.075 }' 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.075 [2024-12-07 05:43:52.389568] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:19.075 "name": "raid_bdev1", 00:16:19.075 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:19.075 "strip_size_kb": 0, 00:16:19.075 "state": "online", 00:16:19.075 "raid_level": "raid1", 00:16:19.075 "superblock": true, 00:16:19.075 "num_base_bdevs": 2, 00:16:19.075 "num_base_bdevs_discovered": 1, 00:16:19.075 "num_base_bdevs_operational": 1, 00:16:19.075 "base_bdevs_list": [ 00:16:19.075 { 00:16:19.075 "name": null, 00:16:19.075 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:19.075 "is_configured": false, 00:16:19.075 "data_offset": 0, 00:16:19.075 "data_size": 7936 00:16:19.075 }, 00:16:19.075 { 00:16:19.075 "name": "BaseBdev2", 00:16:19.075 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:19.075 "is_configured": true, 00:16:19.075 "data_offset": 256, 00:16:19.075 "data_size": 7936 00:16:19.075 } 00:16:19.075 ] 00:16:19.075 }' 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:19.075 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.645 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:19.645 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.645 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.645 [2024-12-07 05:43:52.844783] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:19.645 [2024-12-07 05:43:52.844991] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:19.645 [2024-12-07 05:43:52.845055] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:19.645 [2024-12-07 05:43:52.845110] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:19.645 [2024-12-07 05:43:52.848587] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:16:19.645 [2024-12-07 05:43:52.850477] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:19.645 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:19.645 05:43:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:20.584 "name": "raid_bdev1", 00:16:20.584 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:20.584 "strip_size_kb": 0, 00:16:20.584 "state": "online", 00:16:20.584 "raid_level": "raid1", 00:16:20.584 "superblock": true, 00:16:20.584 "num_base_bdevs": 2, 00:16:20.584 "num_base_bdevs_discovered": 2, 00:16:20.584 "num_base_bdevs_operational": 2, 00:16:20.584 "process": { 00:16:20.584 "type": "rebuild", 00:16:20.584 "target": "spare", 00:16:20.584 "progress": { 00:16:20.584 "blocks": 2560, 00:16:20.584 "percent": 32 00:16:20.584 } 00:16:20.584 }, 00:16:20.584 "base_bdevs_list": [ 00:16:20.584 { 00:16:20.584 "name": "spare", 00:16:20.584 "uuid": "b6f4fc60-d07d-58b9-9a25-1b58d45d7306", 00:16:20.584 "is_configured": true, 00:16:20.584 "data_offset": 256, 00:16:20.584 "data_size": 7936 00:16:20.584 }, 00:16:20.584 { 00:16:20.584 "name": "BaseBdev2", 00:16:20.584 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:20.584 "is_configured": true, 00:16:20.584 "data_offset": 256, 00:16:20.584 "data_size": 7936 00:16:20.584 } 00:16:20.584 ] 00:16:20.584 }' 00:16:20.584 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:20.843 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:20.843 05:43:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:20.843 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:20.843 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:20.843 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.843 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.843 [2024-12-07 05:43:54.017183] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:20.843 [2024-12-07 05:43:54.054438] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:20.843 [2024-12-07 05:43:54.054537] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:20.843 [2024-12-07 05:43:54.054592] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:20.843 [2024-12-07 05:43:54.054613] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:20.843 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.843 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:20.843 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:20.843 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:20.843 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:20.843 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:20.843 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:20.844 "name": "raid_bdev1", 00:16:20.844 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:20.844 "strip_size_kb": 0, 00:16:20.844 "state": "online", 00:16:20.844 "raid_level": "raid1", 00:16:20.844 "superblock": true, 00:16:20.844 "num_base_bdevs": 2, 00:16:20.844 "num_base_bdevs_discovered": 1, 00:16:20.844 "num_base_bdevs_operational": 1, 00:16:20.844 "base_bdevs_list": [ 00:16:20.844 { 00:16:20.844 "name": null, 00:16:20.844 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:20.844 "is_configured": false, 00:16:20.844 "data_offset": 0, 00:16:20.844 "data_size": 7936 00:16:20.844 }, 00:16:20.844 { 00:16:20.844 "name": "BaseBdev2", 00:16:20.844 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:20.844 "is_configured": true, 00:16:20.844 "data_offset": 256, 00:16:20.844 "data_size": 7936 00:16:20.844 } 00:16:20.844 ] 00:16:20.844 }' 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:20.844 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.413 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:21.413 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:21.413 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.413 [2024-12-07 05:43:54.505652] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:21.413 [2024-12-07 05:43:54.505750] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:21.413 [2024-12-07 05:43:54.505795] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:16:21.413 [2024-12-07 05:43:54.505834] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:21.413 [2024-12-07 05:43:54.506039] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:21.413 [2024-12-07 05:43:54.506086] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:21.413 [2024-12-07 05:43:54.506167] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:21.413 [2024-12-07 05:43:54.506202] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:21.413 [2024-12-07 05:43:54.506241] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:21.413 [2024-12-07 05:43:54.506341] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:21.413 [2024-12-07 05:43:54.509832] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:16:21.413 spare 00:16:21.413 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:21.413 05:43:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:21.413 [2024-12-07 05:43:54.511695] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:22.350 "name": "raid_bdev1", 00:16:22.350 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:22.350 "strip_size_kb": 0, 00:16:22.350 "state": "online", 00:16:22.350 "raid_level": "raid1", 00:16:22.350 "superblock": true, 00:16:22.350 "num_base_bdevs": 2, 00:16:22.350 "num_base_bdevs_discovered": 2, 00:16:22.350 "num_base_bdevs_operational": 2, 00:16:22.350 "process": { 00:16:22.350 "type": "rebuild", 00:16:22.350 "target": "spare", 00:16:22.350 "progress": { 00:16:22.350 "blocks": 2560, 00:16:22.350 "percent": 32 00:16:22.350 } 00:16:22.350 }, 00:16:22.350 "base_bdevs_list": [ 00:16:22.350 { 00:16:22.350 "name": "spare", 00:16:22.350 "uuid": "b6f4fc60-d07d-58b9-9a25-1b58d45d7306", 00:16:22.350 "is_configured": true, 00:16:22.350 "data_offset": 256, 00:16:22.350 "data_size": 7936 00:16:22.350 }, 00:16:22.350 { 00:16:22.350 "name": "BaseBdev2", 00:16:22.350 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:22.350 "is_configured": true, 00:16:22.350 "data_offset": 256, 00:16:22.350 "data_size": 7936 00:16:22.350 } 00:16:22.350 ] 00:16:22.350 }' 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:22.350 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.350 [2024-12-07 05:43:55.652501] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:22.350 [2024-12-07 05:43:55.715557] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:22.350 [2024-12-07 05:43:55.715697] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:22.350 [2024-12-07 05:43:55.715733] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:22.350 [2024-12-07 05:43:55.715755] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:22.609 "name": "raid_bdev1", 00:16:22.609 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:22.609 "strip_size_kb": 0, 00:16:22.609 "state": "online", 00:16:22.609 "raid_level": "raid1", 00:16:22.609 "superblock": true, 00:16:22.609 "num_base_bdevs": 2, 00:16:22.609 "num_base_bdevs_discovered": 1, 00:16:22.609 "num_base_bdevs_operational": 1, 00:16:22.609 "base_bdevs_list": [ 00:16:22.609 { 00:16:22.609 "name": null, 00:16:22.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:22.609 "is_configured": false, 00:16:22.609 "data_offset": 0, 00:16:22.609 "data_size": 7936 00:16:22.609 }, 00:16:22.609 { 00:16:22.609 "name": "BaseBdev2", 00:16:22.609 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:22.609 "is_configured": true, 00:16:22.609 "data_offset": 256, 00:16:22.609 "data_size": 7936 00:16:22.609 } 00:16:22.609 ] 00:16:22.609 }' 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:22.609 05:43:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:22.868 "name": "raid_bdev1", 00:16:22.868 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:22.868 "strip_size_kb": 0, 00:16:22.868 "state": "online", 00:16:22.868 "raid_level": "raid1", 00:16:22.868 "superblock": true, 00:16:22.868 "num_base_bdevs": 2, 00:16:22.868 "num_base_bdevs_discovered": 1, 00:16:22.868 "num_base_bdevs_operational": 1, 00:16:22.868 "base_bdevs_list": [ 00:16:22.868 { 00:16:22.868 "name": null, 00:16:22.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:22.868 "is_configured": false, 00:16:22.868 "data_offset": 0, 00:16:22.868 "data_size": 7936 00:16:22.868 }, 00:16:22.868 { 00:16:22.868 "name": "BaseBdev2", 00:16:22.868 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:22.868 "is_configured": true, 00:16:22.868 "data_offset": 256, 00:16:22.868 "data_size": 7936 00:16:22.868 } 00:16:22.868 ] 00:16:22.868 }' 00:16:22.868 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:23.127 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:23.127 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:23.127 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:23.127 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:23.127 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.127 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.127 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.127 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:23.127 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.127 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.127 [2024-12-07 05:43:56.306675] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:23.127 [2024-12-07 05:43:56.306767] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.127 [2024-12-07 05:43:56.306794] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:16:23.127 [2024-12-07 05:43:56.306805] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.127 [2024-12-07 05:43:56.306967] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.127 [2024-12-07 05:43:56.306984] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:23.127 [2024-12-07 05:43:56.307028] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:23.127 [2024-12-07 05:43:56.307043] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:23.128 [2024-12-07 05:43:56.307050] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:23.128 [2024-12-07 05:43:56.307072] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:23.128 BaseBdev1 00:16:23.128 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.128 05:43:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.061 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.062 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.062 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.062 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.062 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:24.062 "name": "raid_bdev1", 00:16:24.062 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:24.062 "strip_size_kb": 0, 00:16:24.062 "state": "online", 00:16:24.062 "raid_level": "raid1", 00:16:24.062 "superblock": true, 00:16:24.062 "num_base_bdevs": 2, 00:16:24.062 "num_base_bdevs_discovered": 1, 00:16:24.062 "num_base_bdevs_operational": 1, 00:16:24.062 "base_bdevs_list": [ 00:16:24.062 { 00:16:24.062 "name": null, 00:16:24.062 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.062 "is_configured": false, 00:16:24.062 "data_offset": 0, 00:16:24.062 "data_size": 7936 00:16:24.062 }, 00:16:24.062 { 00:16:24.062 "name": "BaseBdev2", 00:16:24.062 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:24.062 "is_configured": true, 00:16:24.062 "data_offset": 256, 00:16:24.062 "data_size": 7936 00:16:24.062 } 00:16:24.062 ] 00:16:24.062 }' 00:16:24.062 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:24.062 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:24.630 "name": "raid_bdev1", 00:16:24.630 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:24.630 "strip_size_kb": 0, 00:16:24.630 "state": "online", 00:16:24.630 "raid_level": "raid1", 00:16:24.630 "superblock": true, 00:16:24.630 "num_base_bdevs": 2, 00:16:24.630 "num_base_bdevs_discovered": 1, 00:16:24.630 "num_base_bdevs_operational": 1, 00:16:24.630 "base_bdevs_list": [ 00:16:24.630 { 00:16:24.630 "name": null, 00:16:24.630 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.630 "is_configured": false, 00:16:24.630 "data_offset": 0, 00:16:24.630 "data_size": 7936 00:16:24.630 }, 00:16:24.630 { 00:16:24.630 "name": "BaseBdev2", 00:16:24.630 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:24.630 "is_configured": true, 00:16:24.630 "data_offset": 256, 00:16:24.630 "data_size": 7936 00:16:24.630 } 00:16:24.630 ] 00:16:24.630 }' 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.630 [2024-12-07 05:43:57.880050] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:24.630 [2024-12-07 05:43:57.880261] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:24.630 [2024-12-07 05:43:57.880319] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:24.630 request: 00:16:24.630 { 00:16:24.630 "base_bdev": "BaseBdev1", 00:16:24.630 "raid_bdev": "raid_bdev1", 00:16:24.630 "method": "bdev_raid_add_base_bdev", 00:16:24.630 "req_id": 1 00:16:24.630 } 00:16:24.630 Got JSON-RPC error response 00:16:24.630 response: 00:16:24.630 { 00:16:24.630 "code": -22, 00:16:24.630 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:24.630 } 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:24.630 05:43:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:25.570 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:25.830 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:25.830 "name": "raid_bdev1", 00:16:25.830 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:25.830 "strip_size_kb": 0, 00:16:25.830 "state": "online", 00:16:25.830 "raid_level": "raid1", 00:16:25.830 "superblock": true, 00:16:25.830 "num_base_bdevs": 2, 00:16:25.830 "num_base_bdevs_discovered": 1, 00:16:25.830 "num_base_bdevs_operational": 1, 00:16:25.830 "base_bdevs_list": [ 00:16:25.830 { 00:16:25.830 "name": null, 00:16:25.830 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:25.830 "is_configured": false, 00:16:25.830 "data_offset": 0, 00:16:25.830 "data_size": 7936 00:16:25.830 }, 00:16:25.830 { 00:16:25.830 "name": "BaseBdev2", 00:16:25.830 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:25.830 "is_configured": true, 00:16:25.830 "data_offset": 256, 00:16:25.830 "data_size": 7936 00:16:25.830 } 00:16:25.830 ] 00:16:25.830 }' 00:16:25.830 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:25.830 05:43:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:26.090 "name": "raid_bdev1", 00:16:26.090 "uuid": "5a374777-0742-4a15-8dd1-7cac458942ca", 00:16:26.090 "strip_size_kb": 0, 00:16:26.090 "state": "online", 00:16:26.090 "raid_level": "raid1", 00:16:26.090 "superblock": true, 00:16:26.090 "num_base_bdevs": 2, 00:16:26.090 "num_base_bdevs_discovered": 1, 00:16:26.090 "num_base_bdevs_operational": 1, 00:16:26.090 "base_bdevs_list": [ 00:16:26.090 { 00:16:26.090 "name": null, 00:16:26.090 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:26.090 "is_configured": false, 00:16:26.090 "data_offset": 0, 00:16:26.090 "data_size": 7936 00:16:26.090 }, 00:16:26.090 { 00:16:26.090 "name": "BaseBdev2", 00:16:26.090 "uuid": "2117abd8-1863-5938-85e1-d1cf4fb9888b", 00:16:26.090 "is_configured": true, 00:16:26.090 "data_offset": 256, 00:16:26.090 "data_size": 7936 00:16:26.090 } 00:16:26.090 ] 00:16:26.090 }' 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 98890 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 98890 ']' 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 98890 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:26.090 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 98890 00:16:26.350 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:26.350 killing process with pid 98890 00:16:26.350 Received shutdown signal, test time was about 60.000000 seconds 00:16:26.350 00:16:26.350 Latency(us) 00:16:26.350 [2024-12-07T05:43:59.718Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:26.350 [2024-12-07T05:43:59.718Z] =================================================================================================================== 00:16:26.350 [2024-12-07T05:43:59.718Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:26.350 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:26.350 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 98890' 00:16:26.350 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 98890 00:16:26.350 [2024-12-07 05:43:59.478906] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:26.350 [2024-12-07 05:43:59.479023] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:26.350 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 98890 00:16:26.350 [2024-12-07 05:43:59.479073] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:26.350 [2024-12-07 05:43:59.479082] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:16:26.350 [2024-12-07 05:43:59.510999] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:26.350 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:16:26.350 00:16:26.350 real 0m15.893s 00:16:26.350 user 0m21.220s 00:16:26.350 sys 0m1.569s 00:16:26.350 ************************************ 00:16:26.351 END TEST raid_rebuild_test_sb_md_interleaved 00:16:26.351 ************************************ 00:16:26.351 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:26.351 05:43:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:26.610 05:43:59 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:16:26.610 05:43:59 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:16:26.610 05:43:59 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 98890 ']' 00:16:26.610 05:43:59 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 98890 00:16:26.610 05:43:59 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:16:26.610 00:16:26.610 real 9m45.893s 00:16:26.610 user 13m56.257s 00:16:26.610 sys 1m41.829s 00:16:26.610 ************************************ 00:16:26.610 END TEST bdev_raid 00:16:26.610 ************************************ 00:16:26.610 05:43:59 bdev_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:26.610 05:43:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:26.610 05:43:59 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:26.610 05:43:59 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:16:26.610 05:43:59 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:26.610 05:43:59 -- common/autotest_common.sh@10 -- # set +x 00:16:26.610 ************************************ 00:16:26.610 START TEST spdkcli_raid 00:16:26.610 ************************************ 00:16:26.610 05:43:59 spdkcli_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:26.871 * Looking for test storage... 00:16:26.871 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:26.871 05:43:59 spdkcli_raid -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:16:26.871 05:43:59 spdkcli_raid -- common/autotest_common.sh@1711 -- # lcov --version 00:16:26.871 05:43:59 spdkcli_raid -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:16:26.871 05:44:00 spdkcli_raid -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:26.871 05:44:00 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:16:26.871 05:44:00 spdkcli_raid -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:26.871 05:44:00 spdkcli_raid -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:16:26.871 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:26.871 --rc genhtml_branch_coverage=1 00:16:26.871 --rc genhtml_function_coverage=1 00:16:26.871 --rc genhtml_legend=1 00:16:26.871 --rc geninfo_all_blocks=1 00:16:26.871 --rc geninfo_unexecuted_blocks=1 00:16:26.871 00:16:26.871 ' 00:16:26.871 05:44:00 spdkcli_raid -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:16:26.871 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:26.871 --rc genhtml_branch_coverage=1 00:16:26.871 --rc genhtml_function_coverage=1 00:16:26.871 --rc genhtml_legend=1 00:16:26.871 --rc geninfo_all_blocks=1 00:16:26.871 --rc geninfo_unexecuted_blocks=1 00:16:26.871 00:16:26.871 ' 00:16:26.871 05:44:00 spdkcli_raid -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:16:26.871 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:26.871 --rc genhtml_branch_coverage=1 00:16:26.871 --rc genhtml_function_coverage=1 00:16:26.871 --rc genhtml_legend=1 00:16:26.871 --rc geninfo_all_blocks=1 00:16:26.871 --rc geninfo_unexecuted_blocks=1 00:16:26.871 00:16:26.871 ' 00:16:26.871 05:44:00 spdkcli_raid -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:16:26.871 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:26.871 --rc genhtml_branch_coverage=1 00:16:26.871 --rc genhtml_function_coverage=1 00:16:26.871 --rc genhtml_legend=1 00:16:26.871 --rc geninfo_all_blocks=1 00:16:26.871 --rc geninfo_unexecuted_blocks=1 00:16:26.871 00:16:26.871 ' 00:16:26.871 05:44:00 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:16:26.872 05:44:00 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:16:26.872 05:44:00 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:26.872 05:44:00 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=99561 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:16:26.872 05:44:00 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 99561 00:16:26.872 05:44:00 spdkcli_raid -- common/autotest_common.sh@835 -- # '[' -z 99561 ']' 00:16:26.872 05:44:00 spdkcli_raid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:26.872 05:44:00 spdkcli_raid -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:26.872 05:44:00 spdkcli_raid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:26.872 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:26.872 05:44:00 spdkcli_raid -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:26.872 05:44:00 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:26.872 [2024-12-07 05:44:00.215197] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:16:26.872 [2024-12-07 05:44:00.215431] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99561 ] 00:16:27.132 [2024-12-07 05:44:00.370131] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:16:27.132 [2024-12-07 05:44:00.395922] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:27.132 [2024-12-07 05:44:00.396013] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:27.701 05:44:01 spdkcli_raid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:27.701 05:44:01 spdkcli_raid -- common/autotest_common.sh@868 -- # return 0 00:16:27.701 05:44:01 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:16:27.701 05:44:01 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:27.701 05:44:01 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:27.960 05:44:01 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:16:27.960 05:44:01 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:27.960 05:44:01 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:27.960 05:44:01 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:16:27.960 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:16:27.960 ' 00:16:29.336 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:16:29.336 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:16:29.594 05:44:02 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:16:29.594 05:44:02 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:29.594 05:44:02 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:29.594 05:44:02 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:16:29.594 05:44:02 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:29.594 05:44:02 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:29.594 05:44:02 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:16:29.594 ' 00:16:30.526 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:16:30.783 05:44:03 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:16:30.783 05:44:03 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:30.783 05:44:03 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:30.783 05:44:03 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:16:30.783 05:44:03 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:30.783 05:44:03 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:30.783 05:44:04 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:16:30.784 05:44:04 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:16:31.348 05:44:04 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:16:31.348 05:44:04 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:16:31.348 05:44:04 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:16:31.348 05:44:04 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:31.348 05:44:04 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:31.348 05:44:04 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:16:31.348 05:44:04 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:31.348 05:44:04 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:31.348 05:44:04 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:16:31.348 ' 00:16:32.279 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:16:32.537 05:44:05 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:16:32.537 05:44:05 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:32.537 05:44:05 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:32.537 05:44:05 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:16:32.537 05:44:05 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:32.537 05:44:05 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:32.537 05:44:05 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:16:32.537 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:16:32.537 ' 00:16:33.913 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:16:33.913 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:16:33.913 05:44:07 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:33.913 05:44:07 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 99561 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 99561 ']' 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 99561 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@959 -- # uname 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 99561 00:16:33.913 killing process with pid 99561 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 99561' 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@973 -- # kill 99561 00:16:33.913 05:44:07 spdkcli_raid -- common/autotest_common.sh@978 -- # wait 99561 00:16:34.502 05:44:07 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:16:34.502 05:44:07 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 99561 ']' 00:16:34.502 05:44:07 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 99561 00:16:34.502 05:44:07 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 99561 ']' 00:16:34.502 05:44:07 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 99561 00:16:34.502 Process with pid 99561 is not found 00:16:34.502 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (99561) - No such process 00:16:34.502 05:44:07 spdkcli_raid -- common/autotest_common.sh@981 -- # echo 'Process with pid 99561 is not found' 00:16:34.502 05:44:07 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:16:34.502 05:44:07 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:16:34.502 05:44:07 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:16:34.502 05:44:07 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:16:34.502 00:16:34.502 real 0m7.780s 00:16:34.502 user 0m16.530s 00:16:34.502 sys 0m1.120s 00:16:34.502 05:44:07 spdkcli_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:34.502 05:44:07 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:34.502 ************************************ 00:16:34.502 END TEST spdkcli_raid 00:16:34.502 ************************************ 00:16:34.502 05:44:07 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:16:34.502 05:44:07 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:16:34.502 05:44:07 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:34.502 05:44:07 -- common/autotest_common.sh@10 -- # set +x 00:16:34.502 ************************************ 00:16:34.502 START TEST blockdev_raid5f 00:16:34.502 ************************************ 00:16:34.502 05:44:07 blockdev_raid5f -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:16:34.502 * Looking for test storage... 00:16:34.502 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:16:34.502 05:44:07 blockdev_raid5f -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:16:34.502 05:44:07 blockdev_raid5f -- common/autotest_common.sh@1711 -- # lcov --version 00:16:34.502 05:44:07 blockdev_raid5f -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:16:34.782 05:44:07 blockdev_raid5f -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:34.782 05:44:07 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:16:34.782 05:44:07 blockdev_raid5f -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:34.782 05:44:07 blockdev_raid5f -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:16:34.782 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:34.782 --rc genhtml_branch_coverage=1 00:16:34.782 --rc genhtml_function_coverage=1 00:16:34.782 --rc genhtml_legend=1 00:16:34.782 --rc geninfo_all_blocks=1 00:16:34.782 --rc geninfo_unexecuted_blocks=1 00:16:34.782 00:16:34.782 ' 00:16:34.782 05:44:07 blockdev_raid5f -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:16:34.782 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:34.782 --rc genhtml_branch_coverage=1 00:16:34.782 --rc genhtml_function_coverage=1 00:16:34.782 --rc genhtml_legend=1 00:16:34.782 --rc geninfo_all_blocks=1 00:16:34.782 --rc geninfo_unexecuted_blocks=1 00:16:34.782 00:16:34.782 ' 00:16:34.782 05:44:07 blockdev_raid5f -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:16:34.782 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:34.782 --rc genhtml_branch_coverage=1 00:16:34.782 --rc genhtml_function_coverage=1 00:16:34.782 --rc genhtml_legend=1 00:16:34.782 --rc geninfo_all_blocks=1 00:16:34.782 --rc geninfo_unexecuted_blocks=1 00:16:34.782 00:16:34.782 ' 00:16:34.782 05:44:07 blockdev_raid5f -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:16:34.782 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:34.782 --rc genhtml_branch_coverage=1 00:16:34.782 --rc genhtml_function_coverage=1 00:16:34.782 --rc genhtml_legend=1 00:16:34.782 --rc geninfo_all_blocks=1 00:16:34.782 --rc geninfo_unexecuted_blocks=1 00:16:34.782 00:16:34.782 ' 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@707 -- # QOS_DEV_1=Malloc_0 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@708 -- # QOS_DEV_2=Null_1 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@709 -- # QOS_RUN_TIME=5 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@711 -- # uname -s 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@711 -- # '[' Linux = Linux ']' 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@713 -- # PRE_RESERVED_MEM=0 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@719 -- # test_type=raid5f 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@720 -- # crypto_device= 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@721 -- # dek= 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@722 -- # env_ctx= 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@723 -- # wait_for_rpc= 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@724 -- # '[' -n '' ']' 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == bdev ]] 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == crypto_* ]] 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@730 -- # start_spdk_tgt 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=99815 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:16:34.782 05:44:07 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 99815 00:16:34.782 05:44:07 blockdev_raid5f -- common/autotest_common.sh@835 -- # '[' -z 99815 ']' 00:16:34.782 05:44:07 blockdev_raid5f -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:34.782 05:44:07 blockdev_raid5f -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:34.783 05:44:07 blockdev_raid5f -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:34.783 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:34.783 05:44:07 blockdev_raid5f -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:34.783 05:44:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:34.783 [2024-12-07 05:44:08.046981] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:16:34.783 [2024-12-07 05:44:08.047200] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99815 ] 00:16:35.065 [2024-12-07 05:44:08.200714] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:35.065 [2024-12-07 05:44:08.225073] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@868 -- # return 0 00:16:35.634 05:44:08 blockdev_raid5f -- bdev/blockdev.sh@731 -- # case "$test_type" in 00:16:35.634 05:44:08 blockdev_raid5f -- bdev/blockdev.sh@763 -- # setup_raid5f_conf 00:16:35.634 05:44:08 blockdev_raid5f -- bdev/blockdev.sh@279 -- # rpc_cmd 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:35.634 Malloc0 00:16:35.634 Malloc1 00:16:35.634 Malloc2 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:35.634 05:44:08 blockdev_raid5f -- bdev/blockdev.sh@774 -- # rpc_cmd bdev_wait_for_examine 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:35.634 05:44:08 blockdev_raid5f -- bdev/blockdev.sh@777 -- # cat 00:16:35.634 05:44:08 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n accel 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:35.634 05:44:08 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n bdev 00:16:35.634 05:44:08 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:35.635 05:44:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:35.635 05:44:08 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:35.635 05:44:08 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n iobuf 00:16:35.635 05:44:08 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:35.635 05:44:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:35.635 05:44:08 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:35.635 05:44:08 blockdev_raid5f -- bdev/blockdev.sh@785 -- # mapfile -t bdevs 00:16:35.635 05:44:08 blockdev_raid5f -- bdev/blockdev.sh@785 -- # jq -r '.[] | select(.claimed == false)' 00:16:35.635 05:44:08 blockdev_raid5f -- bdev/blockdev.sh@785 -- # rpc_cmd bdev_get_bdevs 00:16:35.635 05:44:08 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:35.635 05:44:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:35.894 05:44:09 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:35.894 05:44:09 blockdev_raid5f -- bdev/blockdev.sh@786 -- # mapfile -t bdevs_name 00:16:35.894 05:44:09 blockdev_raid5f -- bdev/blockdev.sh@786 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "81e53c1e-5aba-4c35-8eac-c03299161919"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "81e53c1e-5aba-4c35-8eac-c03299161919",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "81e53c1e-5aba-4c35-8eac-c03299161919",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "86628f5d-6af2-4c77-a4fb-ec82b99a0e90",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "8722656a-239a-438a-9e8c-c75b520e4d8c",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "01b5639c-0b33-449f-b76f-03198d561985",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:16:35.894 05:44:09 blockdev_raid5f -- bdev/blockdev.sh@786 -- # jq -r .name 00:16:35.894 05:44:09 blockdev_raid5f -- bdev/blockdev.sh@787 -- # bdev_list=("${bdevs_name[@]}") 00:16:35.894 05:44:09 blockdev_raid5f -- bdev/blockdev.sh@789 -- # hello_world_bdev=raid5f 00:16:35.894 05:44:09 blockdev_raid5f -- bdev/blockdev.sh@790 -- # trap - SIGINT SIGTERM EXIT 00:16:35.894 05:44:09 blockdev_raid5f -- bdev/blockdev.sh@791 -- # killprocess 99815 00:16:35.894 05:44:09 blockdev_raid5f -- common/autotest_common.sh@954 -- # '[' -z 99815 ']' 00:16:35.894 05:44:09 blockdev_raid5f -- common/autotest_common.sh@958 -- # kill -0 99815 00:16:35.894 05:44:09 blockdev_raid5f -- common/autotest_common.sh@959 -- # uname 00:16:35.894 05:44:09 blockdev_raid5f -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:35.894 05:44:09 blockdev_raid5f -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 99815 00:16:35.894 killing process with pid 99815 00:16:35.894 05:44:09 blockdev_raid5f -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:35.894 05:44:09 blockdev_raid5f -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:35.894 05:44:09 blockdev_raid5f -- common/autotest_common.sh@972 -- # echo 'killing process with pid 99815' 00:16:35.894 05:44:09 blockdev_raid5f -- common/autotest_common.sh@973 -- # kill 99815 00:16:35.894 05:44:09 blockdev_raid5f -- common/autotest_common.sh@978 -- # wait 99815 00:16:36.464 05:44:09 blockdev_raid5f -- bdev/blockdev.sh@795 -- # trap cleanup SIGINT SIGTERM EXIT 00:16:36.464 05:44:09 blockdev_raid5f -- bdev/blockdev.sh@797 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:16:36.464 05:44:09 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:36.464 05:44:09 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:36.464 05:44:09 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:36.464 ************************************ 00:16:36.464 START TEST bdev_hello_world 00:16:36.464 ************************************ 00:16:36.464 05:44:09 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:16:36.464 [2024-12-07 05:44:09.620943] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:16:36.464 [2024-12-07 05:44:09.621072] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99857 ] 00:16:36.464 [2024-12-07 05:44:09.774352] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:36.464 [2024-12-07 05:44:09.798827] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:36.723 [2024-12-07 05:44:09.975816] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:16:36.724 [2024-12-07 05:44:09.975860] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:16:36.724 [2024-12-07 05:44:09.975876] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:16:36.724 [2024-12-07 05:44:09.976158] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:16:36.724 [2024-12-07 05:44:09.976283] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:16:36.724 [2024-12-07 05:44:09.976301] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:16:36.724 [2024-12-07 05:44:09.976350] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:16:36.724 00:16:36.724 [2024-12-07 05:44:09.976397] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:16:36.984 00:16:36.984 real 0m0.661s 00:16:36.984 user 0m0.362s 00:16:36.984 sys 0m0.193s 00:16:36.984 05:44:10 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:36.984 05:44:10 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:16:36.984 ************************************ 00:16:36.984 END TEST bdev_hello_world 00:16:36.984 ************************************ 00:16:36.984 05:44:10 blockdev_raid5f -- bdev/blockdev.sh@798 -- # run_test bdev_bounds bdev_bounds '' 00:16:36.984 05:44:10 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:16:36.984 05:44:10 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:36.984 05:44:10 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:36.984 ************************************ 00:16:36.984 START TEST bdev_bounds 00:16:36.984 ************************************ 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1129 -- # bdev_bounds '' 00:16:36.984 Process bdevio pid: 99882 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@289 -- # bdevio_pid=99882 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@288 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # echo 'Process bdevio pid: 99882' 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # waitforlisten 99882 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@835 -- # '[' -z 99882 ']' 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:36.984 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:36.984 05:44:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:16:36.984 [2024-12-07 05:44:10.345641] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:16:36.984 [2024-12-07 05:44:10.345856] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99882 ] 00:16:37.244 [2024-12-07 05:44:10.501815] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:16:37.244 [2024-12-07 05:44:10.528767] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:37.244 [2024-12-07 05:44:10.528835] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:37.244 [2024-12-07 05:44:10.528971] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@868 -- # return 0 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:16:38.183 I/O targets: 00:16:38.183 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:16:38.183 00:16:38.183 00:16:38.183 CUnit - A unit testing framework for C - Version 2.1-3 00:16:38.183 http://cunit.sourceforge.net/ 00:16:38.183 00:16:38.183 00:16:38.183 Suite: bdevio tests on: raid5f 00:16:38.183 Test: blockdev write read block ...passed 00:16:38.183 Test: blockdev write zeroes read block ...passed 00:16:38.183 Test: blockdev write zeroes read no split ...passed 00:16:38.183 Test: blockdev write zeroes read split ...passed 00:16:38.183 Test: blockdev write zeroes read split partial ...passed 00:16:38.183 Test: blockdev reset ...passed 00:16:38.183 Test: blockdev write read 8 blocks ...passed 00:16:38.183 Test: blockdev write read size > 128k ...passed 00:16:38.183 Test: blockdev write read invalid size ...passed 00:16:38.183 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:16:38.183 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:16:38.183 Test: blockdev write read max offset ...passed 00:16:38.183 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:16:38.183 Test: blockdev writev readv 8 blocks ...passed 00:16:38.183 Test: blockdev writev readv 30 x 1block ...passed 00:16:38.183 Test: blockdev writev readv block ...passed 00:16:38.183 Test: blockdev writev readv size > 128k ...passed 00:16:38.183 Test: blockdev writev readv size > 128k in two iovs ...passed 00:16:38.183 Test: blockdev comparev and writev ...passed 00:16:38.183 Test: blockdev nvme passthru rw ...passed 00:16:38.183 Test: blockdev nvme passthru vendor specific ...passed 00:16:38.183 Test: blockdev nvme admin passthru ...passed 00:16:38.183 Test: blockdev copy ...passed 00:16:38.183 00:16:38.183 Run Summary: Type Total Ran Passed Failed Inactive 00:16:38.183 suites 1 1 n/a 0 0 00:16:38.183 tests 23 23 23 0 0 00:16:38.183 asserts 130 130 130 0 n/a 00:16:38.183 00:16:38.183 Elapsed time = 0.307 seconds 00:16:38.183 0 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # killprocess 99882 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@954 -- # '[' -z 99882 ']' 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@958 -- # kill -0 99882 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # uname 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 99882 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@972 -- # echo 'killing process with pid 99882' 00:16:38.183 killing process with pid 99882 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@973 -- # kill 99882 00:16:38.183 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@978 -- # wait 99882 00:16:38.444 05:44:11 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # trap - SIGINT SIGTERM EXIT 00:16:38.444 00:16:38.444 real 0m1.460s 00:16:38.444 user 0m3.621s 00:16:38.444 sys 0m0.345s 00:16:38.444 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:38.444 05:44:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:16:38.444 ************************************ 00:16:38.444 END TEST bdev_bounds 00:16:38.444 ************************************ 00:16:38.444 05:44:11 blockdev_raid5f -- bdev/blockdev.sh@799 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:16:38.444 05:44:11 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:16:38.444 05:44:11 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:38.444 05:44:11 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:38.444 ************************************ 00:16:38.444 START TEST bdev_nbd 00:16:38.444 ************************************ 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1129 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # uname -s 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # [[ Linux == Linux ]] 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@302 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # bdev_all=('raid5f') 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local bdev_all 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_num=1 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@308 -- # [[ -e /sys/module/nbd ]] 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # local nbd_all 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@311 -- # bdev_num=1 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # nbd_list=('/dev/nbd0') 00:16:38.444 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # local nbd_list 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # bdev_list=('raid5f') 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # local bdev_list 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@317 -- # nbd_pid=99931 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # waitforlisten 99931 /var/tmp/spdk-nbd.sock 00:16:38.704 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@835 -- # '[' -z 99931 ']' 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:38.704 05:44:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:16:38.704 [2024-12-07 05:44:11.886873] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:16:38.704 [2024-12-07 05:44:11.887017] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:38.704 [2024-12-07 05:44:12.042277] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:38.704 [2024-12-07 05:44:12.066555] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # return 0 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:39.644 1+0 records in 00:16:39.644 1+0 records out 00:16:39.644 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000595133 s, 6.9 MB/s 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:16:39.644 05:44:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:39.904 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:16:39.904 { 00:16:39.904 "nbd_device": "/dev/nbd0", 00:16:39.904 "bdev_name": "raid5f" 00:16:39.904 } 00:16:39.904 ]' 00:16:39.904 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:16:39.904 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:16:39.904 { 00:16:39.904 "nbd_device": "/dev/nbd0", 00:16:39.904 "bdev_name": "raid5f" 00:16:39.904 } 00:16:39.904 ]' 00:16:39.904 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:16:39.904 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:39.904 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:39.904 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:39.904 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:39.904 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:39.904 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:39.904 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:40.164 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:40.164 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:40.164 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:40.164 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:40.164 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:40.164 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:40.164 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:40.164 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:40.164 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:40.164 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:40.164 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:40.425 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:16:40.686 /dev/nbd0 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:40.686 1+0 records in 00:16:40.686 1+0 records out 00:16:40.686 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000307883 s, 13.3 MB/s 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:40.686 05:44:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:16:40.946 { 00:16:40.946 "nbd_device": "/dev/nbd0", 00:16:40.946 "bdev_name": "raid5f" 00:16:40.946 } 00:16:40.946 ]' 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:16:40.946 { 00:16:40.946 "nbd_device": "/dev/nbd0", 00:16:40.946 "bdev_name": "raid5f" 00:16:40.946 } 00:16:40.946 ]' 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:16:40.946 256+0 records in 00:16:40.946 256+0 records out 00:16:40.946 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0141664 s, 74.0 MB/s 00:16:40.946 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:16:40.947 256+0 records in 00:16:40.947 256+0 records out 00:16:40.947 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0284645 s, 36.8 MB/s 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:40.947 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:41.206 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:41.207 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:41.207 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:41.207 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:41.207 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:41.207 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:41.207 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:41.207 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:41.207 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:41.207 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:41.207 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:16:41.466 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:16:41.738 malloc_lvol_verify 00:16:41.738 05:44:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:16:41.738 d7f745f0-e89b-43cb-af2d-8d4e97f7e7c9 00:16:41.738 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:16:41.997 955f7c37-06e8-4072-9306-f20e7ba69888 00:16:41.997 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:16:42.256 /dev/nbd0 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:16:42.256 mke2fs 1.47.0 (5-Feb-2023) 00:16:42.256 Discarding device blocks: 0/4096 done 00:16:42.256 Creating filesystem with 4096 1k blocks and 1024 inodes 00:16:42.256 00:16:42.256 Allocating group tables: 0/1 done 00:16:42.256 Writing inode tables: 0/1 done 00:16:42.256 Creating journal (1024 blocks): done 00:16:42.256 Writing superblocks and filesystem accounting information: 0/1 done 00:16:42.256 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:42.256 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # killprocess 99931 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@954 -- # '[' -z 99931 ']' 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@958 -- # kill -0 99931 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # uname 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 99931 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@972 -- # echo 'killing process with pid 99931' 00:16:42.515 killing process with pid 99931 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@973 -- # kill 99931 00:16:42.515 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@978 -- # wait 99931 00:16:42.774 05:44:15 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@326 -- # trap - SIGINT SIGTERM EXIT 00:16:42.774 00:16:42.774 real 0m4.191s 00:16:42.774 user 0m6.138s 00:16:42.774 sys 0m1.159s 00:16:42.774 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:42.774 05:44:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:16:42.774 ************************************ 00:16:42.774 END TEST bdev_nbd 00:16:42.774 ************************************ 00:16:42.774 05:44:16 blockdev_raid5f -- bdev/blockdev.sh@800 -- # [[ y == y ]] 00:16:42.774 05:44:16 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = nvme ']' 00:16:42.774 05:44:16 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = gpt ']' 00:16:42.774 05:44:16 blockdev_raid5f -- bdev/blockdev.sh@805 -- # run_test bdev_fio fio_test_suite '' 00:16:42.774 05:44:16 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:16:42.774 05:44:16 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:42.774 05:44:16 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:42.774 ************************************ 00:16:42.774 START TEST bdev_fio 00:16:42.774 ************************************ 00:16:42.774 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1129 -- # fio_test_suite '' 00:16:42.774 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@330 -- # local env_context 00:16:42.774 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@334 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:16:42.774 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:16:42.774 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@335 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:16:42.774 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # echo '' 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # sed s/--env-context=// 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # env_context= 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@339 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=verify 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type=AIO 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z verify ']' 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' verify == verify ']' 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1318 -- # cat 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1327 -- # '[' AIO == AIO ']' 00:16:42.775 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # /usr/src/fio/fio --version 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1329 -- # echo serialize_overlap=1 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # for b in "${bdevs_name[@]}" 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # echo '[job_raid5f]' 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # echo filename=raid5f 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@346 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1105 -- # '[' 11 -le 1 ']' 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:16:43.035 ************************************ 00:16:43.035 START TEST bdev_fio_rw_verify 00:16:43.035 ************************************ 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1129 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1360 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local sanitizers 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # shift 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # local asan_lib= 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # grep libasan 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # asan_lib=/usr/lib64/libasan.so.8 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1350 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1351 -- # break 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:16:43.035 05:44:16 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:43.296 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:16:43.296 fio-3.35 00:16:43.296 Starting 1 thread 00:16:55.510 00:16:55.510 job_raid5f: (groupid=0, jobs=1): err= 0: pid=100119: Sat Dec 7 05:44:26 2024 00:16:55.510 read: IOPS=11.3k, BW=44.3MiB/s (46.5MB/s)(443MiB/10001msec) 00:16:55.510 slat (nsec): min=18900, max=75429, avg=21407.88, stdev=2443.36 00:16:55.510 clat (usec): min=13, max=370, avg=143.42, stdev=50.84 00:16:55.510 lat (usec): min=35, max=396, avg=164.83, stdev=51.18 00:16:55.510 clat percentiles (usec): 00:16:55.510 | 50.000th=[ 145], 99.000th=[ 241], 99.900th=[ 269], 99.990th=[ 302], 00:16:55.510 | 99.999th=[ 338] 00:16:55.510 write: IOPS=11.9k, BW=46.3MiB/s (48.6MB/s)(458MiB/9882msec); 0 zone resets 00:16:55.510 slat (usec): min=7, max=265, avg=17.62, stdev= 4.05 00:16:55.510 clat (usec): min=61, max=1839, avg=322.10, stdev=47.14 00:16:55.510 lat (usec): min=77, max=2105, avg=339.72, stdev=48.43 00:16:55.510 clat percentiles (usec): 00:16:55.510 | 50.000th=[ 322], 99.000th=[ 424], 99.900th=[ 676], 99.990th=[ 1467], 00:16:55.510 | 99.999th=[ 1745] 00:16:55.510 bw ( KiB/s): min=43432, max=50440, per=99.19%, avg=47064.84, stdev=1825.24, samples=19 00:16:55.510 iops : min=10858, max=12610, avg=11766.21, stdev=456.31, samples=19 00:16:55.510 lat (usec) : 20=0.01%, 50=0.01%, 100=11.78%, 250=39.33%, 500=48.80% 00:16:55.510 lat (usec) : 750=0.06%, 1000=0.02% 00:16:55.510 lat (msec) : 2=0.02% 00:16:55.510 cpu : usr=98.88%, sys=0.45%, ctx=23, majf=0, minf=12471 00:16:55.510 IO depths : 1=7.7%, 2=19.9%, 4=55.1%, 8=17.2%, 16=0.0%, 32=0.0%, >=64=0.0% 00:16:55.510 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:16:55.510 complete : 0=0.0%, 4=90.0%, 8=10.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:16:55.510 issued rwts: total=113472,117220,0,0 short=0,0,0,0 dropped=0,0,0,0 00:16:55.510 latency : target=0, window=0, percentile=100.00%, depth=8 00:16:55.510 00:16:55.510 Run status group 0 (all jobs): 00:16:55.510 READ: bw=44.3MiB/s (46.5MB/s), 44.3MiB/s-44.3MiB/s (46.5MB/s-46.5MB/s), io=443MiB (465MB), run=10001-10001msec 00:16:55.510 WRITE: bw=46.3MiB/s (48.6MB/s), 46.3MiB/s-46.3MiB/s (48.6MB/s-48.6MB/s), io=458MiB (480MB), run=9882-9882msec 00:16:55.510 ----------------------------------------------------- 00:16:55.510 Suppressions used: 00:16:55.510 count bytes template 00:16:55.510 1 7 /usr/src/fio/parse.c 00:16:55.510 221 21216 /usr/src/fio/iolog.c 00:16:55.510 1 8 libtcmalloc_minimal.so 00:16:55.510 1 904 libcrypto.so 00:16:55.510 ----------------------------------------------------- 00:16:55.510 00:16:55.510 00:16:55.510 real 0m11.185s 00:16:55.510 user 0m11.675s 00:16:55.510 sys 0m0.578s 00:16:55.510 05:44:27 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:55.510 05:44:27 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:16:55.510 ************************************ 00:16:55.510 END TEST bdev_fio_rw_verify 00:16:55.510 ************************************ 00:16:55.510 05:44:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@349 -- # rm -f 00:16:55.510 05:44:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:55.510 05:44:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@353 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:16:55.510 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:55.510 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=trim 00:16:55.510 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type= 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z trim ']' 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' trim == verify ']' 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1332 -- # '[' trim == trim ']' 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1333 -- # echo rw=trimwrite 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "81e53c1e-5aba-4c35-8eac-c03299161919"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "81e53c1e-5aba-4c35-8eac-c03299161919",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "81e53c1e-5aba-4c35-8eac-c03299161919",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "86628f5d-6af2-4c77-a4fb-ec82b99a0e90",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "8722656a-239a-438a-9e8c-c75b520e4d8c",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "01b5639c-0b33-449f-b76f-03198d561985",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # [[ -n '' ]] 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@360 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:55.511 /home/vagrant/spdk_repo/spdk 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@361 -- # popd 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # trap - SIGINT SIGTERM EXIT 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # return 0 00:16:55.511 00:16:55.511 real 0m11.443s 00:16:55.511 user 0m11.794s 00:16:55.511 sys 0m0.694s 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:55.511 05:44:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:16:55.511 ************************************ 00:16:55.511 END TEST bdev_fio 00:16:55.511 ************************************ 00:16:55.511 05:44:27 blockdev_raid5f -- bdev/blockdev.sh@812 -- # trap cleanup SIGINT SIGTERM EXIT 00:16:55.511 05:44:27 blockdev_raid5f -- bdev/blockdev.sh@814 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:16:55.511 05:44:27 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:16:55.511 05:44:27 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:55.511 05:44:27 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:55.511 ************************************ 00:16:55.511 START TEST bdev_verify 00:16:55.511 ************************************ 00:16:55.511 05:44:27 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:16:55.511 [2024-12-07 05:44:27.634368] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:16:55.511 [2024-12-07 05:44:27.634943] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100269 ] 00:16:55.511 [2024-12-07 05:44:27.790608] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:16:55.511 [2024-12-07 05:44:27.816669] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:55.511 [2024-12-07 05:44:27.816791] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:55.511 Running I/O for 5 seconds... 00:16:56.712 16042.00 IOPS, 62.66 MiB/s [2024-12-07T05:44:31.021Z] 16831.00 IOPS, 65.75 MiB/s [2024-12-07T05:44:32.397Z] 16983.67 IOPS, 66.34 MiB/s [2024-12-07T05:44:33.336Z] 16834.00 IOPS, 65.76 MiB/s [2024-12-07T05:44:33.336Z] 16700.60 IOPS, 65.24 MiB/s 00:16:59.968 Latency(us) 00:16:59.968 [2024-12-07T05:44:33.336Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:59.968 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:16:59.968 Verification LBA range: start 0x0 length 0x2000 00:16:59.968 raid5f : 5.01 8341.43 32.58 0.00 0.00 22892.00 176.18 22322.31 00:16:59.968 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:16:59.968 Verification LBA range: start 0x2000 length 0x2000 00:16:59.968 raid5f : 5.01 8346.68 32.60 0.00 0.00 23066.36 1416.61 22093.36 00:16:59.968 [2024-12-07T05:44:33.336Z] =================================================================================================================== 00:16:59.968 [2024-12-07T05:44:33.336Z] Total : 16688.11 65.19 0.00 0.00 22979.18 176.18 22322.31 00:16:59.968 00:16:59.968 real 0m5.685s 00:16:59.968 user 0m10.644s 00:16:59.968 sys 0m0.211s 00:16:59.968 05:44:33 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:59.968 05:44:33 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:16:59.968 ************************************ 00:16:59.968 END TEST bdev_verify 00:16:59.968 ************************************ 00:16:59.968 05:44:33 blockdev_raid5f -- bdev/blockdev.sh@815 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:16:59.968 05:44:33 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:16:59.968 05:44:33 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:59.969 05:44:33 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:59.969 ************************************ 00:16:59.969 START TEST bdev_verify_big_io 00:16:59.969 ************************************ 00:16:59.969 05:44:33 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:00.228 [2024-12-07 05:44:33.391979] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:17:00.228 [2024-12-07 05:44:33.392098] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100351 ] 00:17:00.228 [2024-12-07 05:44:33.548007] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:00.228 [2024-12-07 05:44:33.574945] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:00.228 [2024-12-07 05:44:33.575051] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:00.488 Running I/O for 5 seconds... 00:17:02.804 760.00 IOPS, 47.50 MiB/s [2024-12-07T05:44:37.111Z] 888.00 IOPS, 55.50 MiB/s [2024-12-07T05:44:38.050Z] 930.67 IOPS, 58.17 MiB/s [2024-12-07T05:44:39.019Z] 952.00 IOPS, 59.50 MiB/s [2024-12-07T05:44:39.019Z] 964.80 IOPS, 60.30 MiB/s 00:17:05.651 Latency(us) 00:17:05.651 [2024-12-07T05:44:39.019Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:05.651 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:17:05.651 Verification LBA range: start 0x0 length 0x200 00:17:05.651 raid5f : 5.13 494.92 30.93 0.00 0.00 6431610.94 136.83 276567.64 00:17:05.651 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:17:05.652 Verification LBA range: start 0x200 length 0x200 00:17:05.652 raid5f : 5.10 497.67 31.10 0.00 0.00 6367592.93 181.55 274736.07 00:17:05.652 [2024-12-07T05:44:39.020Z] =================================================================================================================== 00:17:05.652 [2024-12-07T05:44:39.020Z] Total : 992.59 62.04 0.00 0.00 6399601.93 136.83 276567.64 00:17:05.910 00:17:05.910 real 0m5.801s 00:17:05.910 user 0m10.876s 00:17:05.910 sys 0m0.214s 00:17:05.910 05:44:39 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:05.910 05:44:39 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:17:05.910 ************************************ 00:17:05.910 END TEST bdev_verify_big_io 00:17:05.910 ************************************ 00:17:05.910 05:44:39 blockdev_raid5f -- bdev/blockdev.sh@816 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:05.910 05:44:39 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:17:05.910 05:44:39 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:05.910 05:44:39 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:05.910 ************************************ 00:17:05.910 START TEST bdev_write_zeroes 00:17:05.910 ************************************ 00:17:05.910 05:44:39 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:05.910 [2024-12-07 05:44:39.259775] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:17:05.910 [2024-12-07 05:44:39.259884] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100431 ] 00:17:06.168 [2024-12-07 05:44:39.413697] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:06.168 [2024-12-07 05:44:39.438204] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:06.427 Running I/O for 1 seconds... 00:17:07.363 29127.00 IOPS, 113.78 MiB/s 00:17:07.363 Latency(us) 00:17:07.363 [2024-12-07T05:44:40.731Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:07.363 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:17:07.363 raid5f : 1.01 29091.62 113.64 0.00 0.00 4387.41 1409.45 6009.85 00:17:07.363 [2024-12-07T05:44:40.731Z] =================================================================================================================== 00:17:07.363 [2024-12-07T05:44:40.731Z] Total : 29091.62 113.64 0.00 0.00 4387.41 1409.45 6009.85 00:17:07.622 00:17:07.622 real 0m1.658s 00:17:07.622 user 0m1.344s 00:17:07.622 sys 0m0.202s 00:17:07.622 05:44:40 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:07.622 05:44:40 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:17:07.622 ************************************ 00:17:07.622 END TEST bdev_write_zeroes 00:17:07.622 ************************************ 00:17:07.622 05:44:40 blockdev_raid5f -- bdev/blockdev.sh@819 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:07.622 05:44:40 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:17:07.622 05:44:40 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:07.622 05:44:40 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:07.622 ************************************ 00:17:07.622 START TEST bdev_json_nonenclosed 00:17:07.622 ************************************ 00:17:07.622 05:44:40 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:07.882 [2024-12-07 05:44:40.993227] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:17:07.882 [2024-12-07 05:44:40.993356] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100469 ] 00:17:07.882 [2024-12-07 05:44:41.149203] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:07.882 [2024-12-07 05:44:41.173998] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:07.882 [2024-12-07 05:44:41.174104] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:17:07.882 [2024-12-07 05:44:41.174120] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:07.882 [2024-12-07 05:44:41.174139] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:08.141 00:17:08.141 real 0m0.339s 00:17:08.141 user 0m0.135s 00:17:08.141 sys 0m0.101s 00:17:08.141 05:44:41 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:08.141 05:44:41 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:17:08.141 ************************************ 00:17:08.141 END TEST bdev_json_nonenclosed 00:17:08.141 ************************************ 00:17:08.141 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@822 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:08.141 05:44:41 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:17:08.141 05:44:41 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:08.141 05:44:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:08.141 ************************************ 00:17:08.141 START TEST bdev_json_nonarray 00:17:08.141 ************************************ 00:17:08.141 05:44:41 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:08.141 [2024-12-07 05:44:41.411079] Starting SPDK v25.01-pre git sha1 a2f5e1c2d / DPDK 22.11.4 initialization... 00:17:08.141 [2024-12-07 05:44:41.411274] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100494 ] 00:17:08.401 [2024-12-07 05:44:41.576430] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:08.401 [2024-12-07 05:44:41.601287] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:08.401 [2024-12-07 05:44:41.601390] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:17:08.401 [2024-12-07 05:44:41.601416] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:08.401 [2024-12-07 05:44:41.601434] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:08.401 00:17:08.401 real 0m0.359s 00:17:08.401 user 0m0.149s 00:17:08.401 sys 0m0.106s 00:17:08.401 05:44:41 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:08.401 05:44:41 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:17:08.401 ************************************ 00:17:08.401 END TEST bdev_json_nonarray 00:17:08.401 ************************************ 00:17:08.401 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@824 -- # [[ raid5f == bdev ]] 00:17:08.401 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@832 -- # [[ raid5f == gpt ]] 00:17:08.401 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@836 -- # [[ raid5f == crypto_sw ]] 00:17:08.401 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@848 -- # trap - SIGINT SIGTERM EXIT 00:17:08.401 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@849 -- # cleanup 00:17:08.401 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:17:08.401 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:08.401 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:17:08.401 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:17:08.401 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:17:08.401 05:44:41 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:17:08.401 ************************************ 00:17:08.401 END TEST blockdev_raid5f 00:17:08.401 ************************************ 00:17:08.401 00:17:08.401 real 0m34.030s 00:17:08.401 user 0m47.039s 00:17:08.401 sys 0m4.195s 00:17:08.401 05:44:41 blockdev_raid5f -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:08.401 05:44:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:08.660 05:44:41 -- spdk/autotest.sh@194 -- # uname -s 00:17:08.660 05:44:41 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:17:08.660 05:44:41 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:17:08.660 05:44:41 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:17:08.660 05:44:41 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@256 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@260 -- # timing_exit lib 00:17:08.660 05:44:41 -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:08.660 05:44:41 -- common/autotest_common.sh@10 -- # set +x 00:17:08.660 05:44:41 -- spdk/autotest.sh@262 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@267 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@276 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@319 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@324 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@333 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@350 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:17:08.660 05:44:41 -- spdk/autotest.sh@359 -- # '[' 0 -eq 1 ']' 00:17:08.661 05:44:41 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:17:08.661 05:44:41 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:17:08.661 05:44:41 -- spdk/autotest.sh@374 -- # [[ 0 -eq 1 ]] 00:17:08.661 05:44:41 -- spdk/autotest.sh@378 -- # [[ '' -eq 1 ]] 00:17:08.661 05:44:41 -- spdk/autotest.sh@385 -- # trap - SIGINT SIGTERM EXIT 00:17:08.661 05:44:41 -- spdk/autotest.sh@387 -- # timing_enter post_cleanup 00:17:08.661 05:44:41 -- common/autotest_common.sh@726 -- # xtrace_disable 00:17:08.661 05:44:41 -- common/autotest_common.sh@10 -- # set +x 00:17:08.661 05:44:41 -- spdk/autotest.sh@388 -- # autotest_cleanup 00:17:08.661 05:44:41 -- common/autotest_common.sh@1396 -- # local autotest_es=0 00:17:08.661 05:44:41 -- common/autotest_common.sh@1397 -- # xtrace_disable 00:17:08.661 05:44:41 -- common/autotest_common.sh@10 -- # set +x 00:17:10.569 INFO: APP EXITING 00:17:10.569 INFO: killing all VMs 00:17:10.569 INFO: killing vhost app 00:17:10.569 INFO: EXIT DONE 00:17:11.153 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:17:11.154 Waiting for block devices as requested 00:17:11.154 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:17:11.154 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:17:12.112 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:17:12.112 Cleaning 00:17:12.112 Removing: /var/run/dpdk/spdk0/config 00:17:12.112 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:17:12.112 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:17:12.112 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:17:12.112 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:17:12.112 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:17:12.112 Removing: /var/run/dpdk/spdk0/hugepage_info 00:17:12.112 Removing: /dev/shm/spdk_tgt_trace.pid68835 00:17:12.112 Removing: /var/run/dpdk/spdk0 00:17:12.112 Removing: /var/run/dpdk/spdk_pid100104 00:17:12.112 Removing: /var/run/dpdk/spdk_pid100269 00:17:12.112 Removing: /var/run/dpdk/spdk_pid100351 00:17:12.112 Removing: /var/run/dpdk/spdk_pid100431 00:17:12.112 Removing: /var/run/dpdk/spdk_pid100469 00:17:12.112 Removing: /var/run/dpdk/spdk_pid100494 00:17:12.112 Removing: /var/run/dpdk/spdk_pid68672 00:17:12.112 Removing: /var/run/dpdk/spdk_pid68835 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69037 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69124 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69153 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69264 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69277 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69465 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69544 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69622 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69718 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69804 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69838 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69875 00:17:12.112 Removing: /var/run/dpdk/spdk_pid69945 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70058 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70482 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70529 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70576 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70592 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70655 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70666 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70737 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70747 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70795 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70813 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70855 00:17:12.112 Removing: /var/run/dpdk/spdk_pid70873 00:17:12.112 Removing: /var/run/dpdk/spdk_pid71000 00:17:12.112 Removing: /var/run/dpdk/spdk_pid71042 00:17:12.112 Removing: /var/run/dpdk/spdk_pid71120 00:17:12.112 Removing: /var/run/dpdk/spdk_pid72285 00:17:12.112 Removing: /var/run/dpdk/spdk_pid72486 00:17:12.112 Removing: /var/run/dpdk/spdk_pid72615 00:17:12.371 Removing: /var/run/dpdk/spdk_pid73214 00:17:12.371 Removing: /var/run/dpdk/spdk_pid73415 00:17:12.371 Removing: /var/run/dpdk/spdk_pid73544 00:17:12.371 Removing: /var/run/dpdk/spdk_pid74144 00:17:12.371 Removing: /var/run/dpdk/spdk_pid74462 00:17:12.371 Removing: /var/run/dpdk/spdk_pid74591 00:17:12.371 Removing: /var/run/dpdk/spdk_pid75921 00:17:12.371 Removing: /var/run/dpdk/spdk_pid76162 00:17:12.371 Removing: /var/run/dpdk/spdk_pid76292 00:17:12.371 Removing: /var/run/dpdk/spdk_pid77616 00:17:12.371 Removing: /var/run/dpdk/spdk_pid77853 00:17:12.371 Removing: /var/run/dpdk/spdk_pid77982 00:17:12.371 Removing: /var/run/dpdk/spdk_pid79317 00:17:12.371 Removing: /var/run/dpdk/spdk_pid79752 00:17:12.371 Removing: /var/run/dpdk/spdk_pid79881 00:17:12.371 Removing: /var/run/dpdk/spdk_pid81311 00:17:12.371 Removing: /var/run/dpdk/spdk_pid81559 00:17:12.371 Removing: /var/run/dpdk/spdk_pid81688 00:17:12.371 Removing: /var/run/dpdk/spdk_pid83118 00:17:12.371 Removing: /var/run/dpdk/spdk_pid83361 00:17:12.371 Removing: /var/run/dpdk/spdk_pid83496 00:17:12.371 Removing: /var/run/dpdk/spdk_pid84932 00:17:12.371 Removing: /var/run/dpdk/spdk_pid85399 00:17:12.371 Removing: /var/run/dpdk/spdk_pid85533 00:17:12.371 Removing: /var/run/dpdk/spdk_pid85660 00:17:12.371 Removing: /var/run/dpdk/spdk_pid86060 00:17:12.371 Removing: /var/run/dpdk/spdk_pid86767 00:17:12.371 Removing: /var/run/dpdk/spdk_pid87129 00:17:12.371 Removing: /var/run/dpdk/spdk_pid87806 00:17:12.371 Removing: /var/run/dpdk/spdk_pid88241 00:17:12.371 Removing: /var/run/dpdk/spdk_pid88979 00:17:12.371 Removing: /var/run/dpdk/spdk_pid89366 00:17:12.371 Removing: /var/run/dpdk/spdk_pid91268 00:17:12.371 Removing: /var/run/dpdk/spdk_pid91690 00:17:12.371 Removing: /var/run/dpdk/spdk_pid92118 00:17:12.371 Removing: /var/run/dpdk/spdk_pid94138 00:17:12.371 Removing: /var/run/dpdk/spdk_pid94607 00:17:12.371 Removing: /var/run/dpdk/spdk_pid95107 00:17:12.371 Removing: /var/run/dpdk/spdk_pid96136 00:17:12.371 Removing: /var/run/dpdk/spdk_pid96446 00:17:12.371 Removing: /var/run/dpdk/spdk_pid97357 00:17:12.371 Removing: /var/run/dpdk/spdk_pid97663 00:17:12.371 Removing: /var/run/dpdk/spdk_pid98578 00:17:12.371 Removing: /var/run/dpdk/spdk_pid98890 00:17:12.371 Removing: /var/run/dpdk/spdk_pid99561 00:17:12.371 Removing: /var/run/dpdk/spdk_pid99815 00:17:12.371 Removing: /var/run/dpdk/spdk_pid99857 00:17:12.371 Removing: /var/run/dpdk/spdk_pid99882 00:17:12.371 Clean 00:17:12.630 05:44:45 -- common/autotest_common.sh@1453 -- # return 0 00:17:12.630 05:44:45 -- spdk/autotest.sh@389 -- # timing_exit post_cleanup 00:17:12.630 05:44:45 -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:12.630 05:44:45 -- common/autotest_common.sh@10 -- # set +x 00:17:12.630 05:44:45 -- spdk/autotest.sh@391 -- # timing_exit autotest 00:17:12.630 05:44:45 -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:12.630 05:44:45 -- common/autotest_common.sh@10 -- # set +x 00:17:12.630 05:44:45 -- spdk/autotest.sh@392 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:17:12.630 05:44:45 -- spdk/autotest.sh@394 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:17:12.630 05:44:45 -- spdk/autotest.sh@394 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:17:12.630 05:44:45 -- spdk/autotest.sh@396 -- # [[ y == y ]] 00:17:12.630 05:44:45 -- spdk/autotest.sh@398 -- # hostname 00:17:12.630 05:44:45 -- spdk/autotest.sh@398 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:17:12.890 geninfo: WARNING: invalid characters removed from testname! 00:17:34.835 05:45:06 -- spdk/autotest.sh@399 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:35.780 05:45:08 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:37.693 05:45:10 -- spdk/autotest.sh@404 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:39.601 05:45:12 -- spdk/autotest.sh@405 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:41.512 05:45:14 -- spdk/autotest.sh@406 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:43.413 05:45:16 -- spdk/autotest.sh@407 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:45.392 05:45:18 -- spdk/autotest.sh@408 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:17:45.392 05:45:18 -- spdk/autorun.sh@1 -- $ timing_finish 00:17:45.392 05:45:18 -- common/autotest_common.sh@738 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/timing.txt ]] 00:17:45.392 05:45:18 -- common/autotest_common.sh@740 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:17:45.392 05:45:18 -- common/autotest_common.sh@741 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:17:45.392 05:45:18 -- common/autotest_common.sh@744 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:17:45.650 + [[ -n 6170 ]] 00:17:45.650 + sudo kill 6170 00:17:45.658 [Pipeline] } 00:17:45.671 [Pipeline] // timeout 00:17:45.676 [Pipeline] } 00:17:45.689 [Pipeline] // stage 00:17:45.694 [Pipeline] } 00:17:45.705 [Pipeline] // catchError 00:17:45.713 [Pipeline] stage 00:17:45.715 [Pipeline] { (Stop VM) 00:17:45.725 [Pipeline] sh 00:17:46.006 + vagrant halt 00:17:48.542 ==> default: Halting domain... 00:17:56.690 [Pipeline] sh 00:17:56.973 + vagrant destroy -f 00:17:59.515 ==> default: Removing domain... 00:17:59.528 [Pipeline] sh 00:17:59.814 + mv output /var/jenkins/workspace/raid-vg-autotest/output 00:17:59.825 [Pipeline] } 00:17:59.844 [Pipeline] // stage 00:17:59.850 [Pipeline] } 00:17:59.867 [Pipeline] // dir 00:17:59.875 [Pipeline] } 00:17:59.890 [Pipeline] // wrap 00:17:59.896 [Pipeline] } 00:17:59.911 [Pipeline] // catchError 00:17:59.920 [Pipeline] stage 00:17:59.922 [Pipeline] { (Epilogue) 00:17:59.936 [Pipeline] sh 00:18:00.226 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:18:05.521 [Pipeline] catchError 00:18:05.523 [Pipeline] { 00:18:05.536 [Pipeline] sh 00:18:05.822 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:18:05.822 Artifacts sizes are good 00:18:05.832 [Pipeline] } 00:18:05.873 [Pipeline] // catchError 00:18:05.886 [Pipeline] archiveArtifacts 00:18:05.893 Archiving artifacts 00:18:05.985 [Pipeline] cleanWs 00:18:05.997 [WS-CLEANUP] Deleting project workspace... 00:18:05.997 [WS-CLEANUP] Deferred wipeout is used... 00:18:06.004 [WS-CLEANUP] done 00:18:06.006 [Pipeline] } 00:18:06.023 [Pipeline] // stage 00:18:06.029 [Pipeline] } 00:18:06.043 [Pipeline] // node 00:18:06.050 [Pipeline] End of Pipeline 00:18:06.153 Finished: SUCCESS